May 15 14:55:59.921802 kernel: Linux version 6.12.20-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Thu May 15 10:42:41 -00 2025 May 15 14:55:59.921844 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 14:55:59.921859 kernel: BIOS-provided physical RAM map: May 15 14:55:59.921870 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable May 15 14:55:59.921881 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved May 15 14:55:59.921904 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved May 15 14:55:59.921916 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable May 15 14:55:59.921933 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved May 15 14:55:59.921948 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 15 14:55:59.921972 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved May 15 14:55:59.921983 kernel: NX (Execute Disable) protection: active May 15 14:55:59.922009 kernel: APIC: Static calls initialized May 15 14:55:59.922020 kernel: SMBIOS 2.8 present. May 15 14:55:59.922032 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 May 15 14:55:59.922049 kernel: DMI: Memory slots populated: 1/1 May 15 14:55:59.922061 kernel: Hypervisor detected: KVM May 15 14:55:59.922077 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 15 14:55:59.922091 kernel: kvm-clock: using sched offset of 4526504569 cycles May 15 14:55:59.922105 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 15 14:55:59.922118 kernel: tsc: Detected 2494.136 MHz processor May 15 14:55:59.922130 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 15 14:55:59.922143 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 15 14:55:59.922155 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 May 15 14:55:59.922172 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs May 15 14:55:59.922184 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 15 14:55:59.922197 kernel: ACPI: Early table checksum verification disabled May 15 14:55:59.922209 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) May 15 14:55:59.922222 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922234 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922246 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922259 kernel: ACPI: FACS 0x000000007FFE0000 000040 May 15 14:55:59.922271 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922287 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922300 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922312 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 15 14:55:59.922324 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] May 15 14:55:59.922336 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] May 15 14:55:59.922348 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] May 15 14:55:59.922361 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] May 15 14:55:59.922373 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] May 15 14:55:59.922396 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] May 15 14:55:59.922409 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] May 15 14:55:59.922423 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 15 14:55:59.922436 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] May 15 14:55:59.922449 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00001000-0x7ffdafff] May 15 14:55:59.922465 kernel: NODE_DATA(0) allocated [mem 0x7ffd3dc0-0x7ffdafff] May 15 14:55:59.922479 kernel: Zone ranges: May 15 14:55:59.922493 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 15 14:55:59.922507 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] May 15 14:55:59.922520 kernel: Normal empty May 15 14:55:59.922533 kernel: Device empty May 15 14:55:59.922547 kernel: Movable zone start for each node May 15 14:55:59.922559 kernel: Early memory node ranges May 15 14:55:59.922571 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] May 15 14:55:59.922582 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] May 15 14:55:59.922598 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] May 15 14:55:59.922608 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 15 14:55:59.922619 kernel: On node 0, zone DMA: 97 pages in unavailable ranges May 15 14:55:59.922631 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges May 15 14:55:59.922643 kernel: ACPI: PM-Timer IO Port: 0x608 May 15 14:55:59.922657 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 15 14:55:59.922674 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 15 14:55:59.922688 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 15 14:55:59.922704 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 15 14:55:59.922722 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 15 14:55:59.922738 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 15 14:55:59.922751 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 15 14:55:59.922763 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 15 14:55:59.922776 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 15 14:55:59.922790 kernel: TSC deadline timer available May 15 14:55:59.922803 kernel: CPU topo: Max. logical packages: 1 May 15 14:55:59.922816 kernel: CPU topo: Max. logical dies: 1 May 15 14:55:59.922829 kernel: CPU topo: Max. dies per package: 1 May 15 14:55:59.922847 kernel: CPU topo: Max. threads per core: 1 May 15 14:55:59.922859 kernel: CPU topo: Num. cores per package: 2 May 15 14:55:59.922873 kernel: CPU topo: Num. threads per package: 2 May 15 14:55:59.922887 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs May 15 14:55:59.922901 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 15 14:55:59.922914 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices May 15 14:55:59.922926 kernel: Booting paravirtualized kernel on KVM May 15 14:55:59.922940 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 15 14:55:59.924992 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 15 14:55:59.925048 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 May 15 14:55:59.925062 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 May 15 14:55:59.925076 kernel: pcpu-alloc: [0] 0 1 May 15 14:55:59.925090 kernel: kvm-guest: PV spinlocks disabled, no host support May 15 14:55:59.925108 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 14:55:59.925122 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 15 14:55:59.925141 kernel: random: crng init done May 15 14:55:59.925156 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 15 14:55:59.925188 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 15 14:55:59.925208 kernel: Fallback order for Node 0: 0 May 15 14:55:59.925227 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524153 May 15 14:55:59.925240 kernel: Policy zone: DMA32 May 15 14:55:59.925255 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 15 14:55:59.925270 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 15 14:55:59.925283 kernel: Kernel/User page tables isolation: enabled May 15 14:55:59.925297 kernel: ftrace: allocating 40065 entries in 157 pages May 15 14:55:59.925311 kernel: ftrace: allocated 157 pages with 5 groups May 15 14:55:59.925325 kernel: Dynamic Preempt: voluntary May 15 14:55:59.925345 kernel: rcu: Preemptible hierarchical RCU implementation. May 15 14:55:59.925360 kernel: rcu: RCU event tracing is enabled. May 15 14:55:59.925373 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 15 14:55:59.925387 kernel: Trampoline variant of Tasks RCU enabled. May 15 14:55:59.925401 kernel: Rude variant of Tasks RCU enabled. May 15 14:55:59.925415 kernel: Tracing variant of Tasks RCU enabled. May 15 14:55:59.925430 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 15 14:55:59.925444 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 15 14:55:59.925459 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 14:55:59.925486 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 14:55:59.925501 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 15 14:55:59.925515 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 15 14:55:59.925529 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 15 14:55:59.925543 kernel: Console: colour VGA+ 80x25 May 15 14:55:59.925556 kernel: printk: legacy console [tty0] enabled May 15 14:55:59.925571 kernel: printk: legacy console [ttyS0] enabled May 15 14:55:59.925584 kernel: ACPI: Core revision 20240827 May 15 14:55:59.925598 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 15 14:55:59.925630 kernel: APIC: Switch to symmetric I/O mode setup May 15 14:55:59.925644 kernel: x2apic enabled May 15 14:55:59.925664 kernel: APIC: Switched APIC routing to: physical x2apic May 15 14:55:59.925678 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 15 14:55:59.925696 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 15 14:55:59.925711 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494136) May 15 14:55:59.925725 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 May 15 14:55:59.925741 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 May 15 14:55:59.925757 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 15 14:55:59.925776 kernel: Spectre V2 : Mitigation: Retpolines May 15 14:55:59.925792 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch May 15 14:55:59.925806 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT May 15 14:55:59.925822 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls May 15 14:55:59.925836 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 15 14:55:59.925852 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 15 14:55:59.925867 kernel: MDS: Mitigation: Clear CPU buffers May 15 14:55:59.925886 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 15 14:55:59.925901 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 15 14:55:59.925916 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 15 14:55:59.925931 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 15 14:55:59.925944 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 15 14:55:59.925974 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 15 14:55:59.925988 kernel: Freeing SMP alternatives memory: 32K May 15 14:55:59.926003 kernel: pid_max: default: 32768 minimum: 301 May 15 14:55:59.926017 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima May 15 14:55:59.926037 kernel: landlock: Up and running. May 15 14:55:59.926052 kernel: SELinux: Initializing. May 15 14:55:59.926067 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 15 14:55:59.926082 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 15 14:55:59.926096 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) May 15 14:55:59.926110 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. May 15 14:55:59.926126 kernel: signal: max sigframe size: 1776 May 15 14:55:59.926140 kernel: rcu: Hierarchical SRCU implementation. May 15 14:55:59.926156 kernel: rcu: Max phase no-delay instances is 400. May 15 14:55:59.926176 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level May 15 14:55:59.926191 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 15 14:55:59.926203 kernel: smp: Bringing up secondary CPUs ... May 15 14:55:59.926218 kernel: smpboot: x86: Booting SMP configuration: May 15 14:55:59.926237 kernel: .... node #0, CPUs: #1 May 15 14:55:59.926252 kernel: smp: Brought up 1 node, 2 CPUs May 15 14:55:59.926266 kernel: smpboot: Total of 2 processors activated (9976.54 BogoMIPS) May 15 14:55:59.926282 kernel: Memory: 1966908K/2096612K available (14336K kernel code, 2438K rwdata, 9944K rodata, 54416K init, 2544K bss, 125140K reserved, 0K cma-reserved) May 15 14:55:59.926296 kernel: devtmpfs: initialized May 15 14:55:59.926317 kernel: x86/mm: Memory block size: 128MB May 15 14:55:59.926331 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 15 14:55:59.926345 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 15 14:55:59.926359 kernel: pinctrl core: initialized pinctrl subsystem May 15 14:55:59.926375 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 15 14:55:59.926391 kernel: audit: initializing netlink subsys (disabled) May 15 14:55:59.926406 kernel: audit: type=2000 audit(1747320956.509:1): state=initialized audit_enabled=0 res=1 May 15 14:55:59.926420 kernel: thermal_sys: Registered thermal governor 'step_wise' May 15 14:55:59.926434 kernel: thermal_sys: Registered thermal governor 'user_space' May 15 14:55:59.926451 kernel: cpuidle: using governor menu May 15 14:55:59.926464 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 15 14:55:59.926476 kernel: dca service started, version 1.12.1 May 15 14:55:59.926489 kernel: PCI: Using configuration type 1 for base access May 15 14:55:59.926503 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 15 14:55:59.926517 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 15 14:55:59.926533 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 15 14:55:59.926547 kernel: ACPI: Added _OSI(Module Device) May 15 14:55:59.926561 kernel: ACPI: Added _OSI(Processor Device) May 15 14:55:59.926581 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 15 14:55:59.926597 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 15 14:55:59.926612 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 15 14:55:59.926629 kernel: ACPI: Interpreter enabled May 15 14:55:59.926643 kernel: ACPI: PM: (supports S0 S5) May 15 14:55:59.926658 kernel: ACPI: Using IOAPIC for interrupt routing May 15 14:55:59.926673 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 15 14:55:59.926688 kernel: PCI: Using E820 reservations for host bridge windows May 15 14:55:59.926703 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 15 14:55:59.926724 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 15 14:55:59.929196 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 15 14:55:59.929411 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 15 14:55:59.929555 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 15 14:55:59.929576 kernel: acpiphp: Slot [3] registered May 15 14:55:59.929591 kernel: acpiphp: Slot [4] registered May 15 14:55:59.929607 kernel: acpiphp: Slot [5] registered May 15 14:55:59.929634 kernel: acpiphp: Slot [6] registered May 15 14:55:59.929648 kernel: acpiphp: Slot [7] registered May 15 14:55:59.929663 kernel: acpiphp: Slot [8] registered May 15 14:55:59.929678 kernel: acpiphp: Slot [9] registered May 15 14:55:59.929693 kernel: acpiphp: Slot [10] registered May 15 14:55:59.929708 kernel: acpiphp: Slot [11] registered May 15 14:55:59.929722 kernel: acpiphp: Slot [12] registered May 15 14:55:59.929747 kernel: acpiphp: Slot [13] registered May 15 14:55:59.929763 kernel: acpiphp: Slot [14] registered May 15 14:55:59.929777 kernel: acpiphp: Slot [15] registered May 15 14:55:59.929797 kernel: acpiphp: Slot [16] registered May 15 14:55:59.929811 kernel: acpiphp: Slot [17] registered May 15 14:55:59.929825 kernel: acpiphp: Slot [18] registered May 15 14:55:59.929841 kernel: acpiphp: Slot [19] registered May 15 14:55:59.929856 kernel: acpiphp: Slot [20] registered May 15 14:55:59.929871 kernel: acpiphp: Slot [21] registered May 15 14:55:59.929886 kernel: acpiphp: Slot [22] registered May 15 14:55:59.929900 kernel: acpiphp: Slot [23] registered May 15 14:55:59.929915 kernel: acpiphp: Slot [24] registered May 15 14:55:59.929936 kernel: acpiphp: Slot [25] registered May 15 14:55:59.929951 kernel: acpiphp: Slot [26] registered May 15 14:55:59.929993 kernel: acpiphp: Slot [27] registered May 15 14:55:59.930007 kernel: acpiphp: Slot [28] registered May 15 14:55:59.931061 kernel: acpiphp: Slot [29] registered May 15 14:55:59.931090 kernel: acpiphp: Slot [30] registered May 15 14:55:59.931106 kernel: acpiphp: Slot [31] registered May 15 14:55:59.931122 kernel: PCI host bridge to bus 0000:00 May 15 14:55:59.931358 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 15 14:55:59.931527 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 15 14:55:59.931662 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 15 14:55:59.931792 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 15 14:55:59.931914 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] May 15 14:55:59.933194 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 15 14:55:59.933409 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint May 15 14:55:59.933577 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint May 15 14:55:59.933772 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 conventional PCI endpoint May 15 14:55:59.933913 kernel: pci 0000:00:01.1: BAR 4 [io 0xc1e0-0xc1ef] May 15 14:55:59.935156 kernel: pci 0000:00:01.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk May 15 14:55:59.935371 kernel: pci 0000:00:01.1: BAR 1 [io 0x03f6]: legacy IDE quirk May 15 14:55:59.935554 kernel: pci 0000:00:01.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk May 15 14:55:59.935695 kernel: pci 0000:00:01.1: BAR 3 [io 0x0376]: legacy IDE quirk May 15 14:55:59.935872 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 conventional PCI endpoint May 15 14:55:59.936913 kernel: pci 0000:00:01.2: BAR 4 [io 0xc180-0xc19f] May 15 14:55:59.937125 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint May 15 14:55:59.937272 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI May 15 14:55:59.937417 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB May 15 14:55:59.937578 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 conventional PCI endpoint May 15 14:55:59.937738 kernel: pci 0000:00:02.0: BAR 0 [mem 0xfe000000-0xfe7fffff pref] May 15 14:55:59.937884 kernel: pci 0000:00:02.0: BAR 2 [mem 0xfe800000-0xfe803fff 64bit pref] May 15 14:55:59.939564 kernel: pci 0000:00:02.0: BAR 4 [mem 0xfebf0000-0xfebf0fff] May 15 14:55:59.939740 kernel: pci 0000:00:02.0: ROM [mem 0xfebe0000-0xfebeffff pref] May 15 14:55:59.939896 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 15 14:55:59.940994 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 15 14:55:59.941171 kernel: pci 0000:00:03.0: BAR 0 [io 0xc1a0-0xc1bf] May 15 14:55:59.941329 kernel: pci 0000:00:03.0: BAR 1 [mem 0xfebf1000-0xfebf1fff] May 15 14:55:59.941475 kernel: pci 0000:00:03.0: BAR 4 [mem 0xfe804000-0xfe807fff 64bit pref] May 15 14:55:59.941645 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 15 14:55:59.941795 kernel: pci 0000:00:04.0: BAR 0 [io 0xc1c0-0xc1df] May 15 14:55:59.941940 kernel: pci 0000:00:04.0: BAR 1 [mem 0xfebf2000-0xfebf2fff] May 15 14:55:59.942123 kernel: pci 0000:00:04.0: BAR 4 [mem 0xfe808000-0xfe80bfff 64bit pref] May 15 14:55:59.942288 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 conventional PCI endpoint May 15 14:55:59.942451 kernel: pci 0000:00:05.0: BAR 0 [io 0xc100-0xc13f] May 15 14:55:59.942597 kernel: pci 0000:00:05.0: BAR 1 [mem 0xfebf3000-0xfebf3fff] May 15 14:55:59.942739 kernel: pci 0000:00:05.0: BAR 4 [mem 0xfe80c000-0xfe80ffff 64bit pref] May 15 14:55:59.942903 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 15 14:55:59.947193 kernel: pci 0000:00:06.0: BAR 0 [io 0xc000-0xc07f] May 15 14:55:59.947403 kernel: pci 0000:00:06.0: BAR 1 [mem 0xfebf4000-0xfebf4fff] May 15 14:55:59.947573 kernel: pci 0000:00:06.0: BAR 4 [mem 0xfe810000-0xfe813fff 64bit pref] May 15 14:55:59.947771 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 15 14:55:59.947919 kernel: pci 0000:00:07.0: BAR 0 [io 0xc080-0xc0ff] May 15 14:55:59.950188 kernel: pci 0000:00:07.0: BAR 1 [mem 0xfebf5000-0xfebf5fff] May 15 14:55:59.950372 kernel: pci 0000:00:07.0: BAR 4 [mem 0xfe814000-0xfe817fff 64bit pref] May 15 14:55:59.950539 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 conventional PCI endpoint May 15 14:55:59.950683 kernel: pci 0000:00:08.0: BAR 0 [io 0xc140-0xc17f] May 15 14:55:59.950836 kernel: pci 0000:00:08.0: BAR 4 [mem 0xfe818000-0xfe81bfff 64bit pref] May 15 14:55:59.950855 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 15 14:55:59.950870 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 15 14:55:59.950885 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 15 14:55:59.950899 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 15 14:55:59.950913 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 15 14:55:59.950929 kernel: iommu: Default domain type: Translated May 15 14:55:59.950946 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 15 14:55:59.951003 kernel: PCI: Using ACPI for IRQ routing May 15 14:55:59.951018 kernel: PCI: pci_cache_line_size set to 64 bytes May 15 14:55:59.951033 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] May 15 14:55:59.951049 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] May 15 14:55:59.951204 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device May 15 14:55:59.951348 kernel: pci 0000:00:02.0: vgaarb: bridge control possible May 15 14:55:59.951538 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 15 14:55:59.951558 kernel: vgaarb: loaded May 15 14:55:59.951574 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 15 14:55:59.951598 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 15 14:55:59.951613 kernel: clocksource: Switched to clocksource kvm-clock May 15 14:55:59.951629 kernel: VFS: Disk quotas dquot_6.6.0 May 15 14:55:59.951646 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 15 14:55:59.951660 kernel: pnp: PnP ACPI init May 15 14:55:59.951674 kernel: pnp: PnP ACPI: found 4 devices May 15 14:55:59.951689 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 15 14:55:59.951705 kernel: NET: Registered PF_INET protocol family May 15 14:55:59.951720 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 15 14:55:59.951740 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 15 14:55:59.951755 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 15 14:55:59.951769 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 15 14:55:59.951785 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) May 15 14:55:59.951801 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 15 14:55:59.951816 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 15 14:55:59.951831 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 15 14:55:59.951846 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 15 14:55:59.951861 kernel: NET: Registered PF_XDP protocol family May 15 14:55:59.952049 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 15 14:55:59.952189 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 15 14:55:59.952315 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 15 14:55:59.952439 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 15 14:55:59.952560 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] May 15 14:55:59.952709 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release May 15 14:55:59.952888 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 15 14:55:59.952919 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 15 14:55:59.957235 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x720 took 32783 usecs May 15 14:55:59.957292 kernel: PCI: CLS 0 bytes, default 64 May 15 14:55:59.957308 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 15 14:55:59.957324 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 15 14:55:59.957339 kernel: Initialise system trusted keyrings May 15 14:55:59.957356 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 15 14:55:59.957371 kernel: Key type asymmetric registered May 15 14:55:59.957386 kernel: Asymmetric key parser 'x509' registered May 15 14:55:59.957412 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 15 14:55:59.957427 kernel: io scheduler mq-deadline registered May 15 14:55:59.957441 kernel: io scheduler kyber registered May 15 14:55:59.957455 kernel: io scheduler bfq registered May 15 14:55:59.957470 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 15 14:55:59.957486 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 May 15 14:55:59.957502 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 15 14:55:59.957516 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 15 14:55:59.957531 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 15 14:55:59.957550 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 15 14:55:59.957564 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 15 14:55:59.957580 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 15 14:55:59.957596 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 15 14:55:59.957806 kernel: rtc_cmos 00:03: RTC can wake from S4 May 15 14:55:59.957831 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 15 14:55:59.957982 kernel: rtc_cmos 00:03: registered as rtc0 May 15 14:55:59.958147 kernel: rtc_cmos 00:03: setting system clock to 2025-05-15T14:55:59 UTC (1747320959) May 15 14:55:59.958292 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram May 15 14:55:59.958310 kernel: intel_pstate: CPU model not supported May 15 14:55:59.958324 kernel: NET: Registered PF_INET6 protocol family May 15 14:55:59.958338 kernel: Segment Routing with IPv6 May 15 14:55:59.958351 kernel: In-situ OAM (IOAM) with IPv6 May 15 14:55:59.958367 kernel: NET: Registered PF_PACKET protocol family May 15 14:55:59.958381 kernel: Key type dns_resolver registered May 15 14:55:59.958397 kernel: IPI shorthand broadcast: enabled May 15 14:55:59.958412 kernel: sched_clock: Marking stable (3587008382, 107409136)->(3714156605, -19739087) May 15 14:55:59.958434 kernel: registered taskstats version 1 May 15 14:55:59.958450 kernel: Loading compiled-in X.509 certificates May 15 14:55:59.958464 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.20-flatcar: 05e05785144663be6df1db78301487421c4773b6' May 15 14:55:59.958479 kernel: Demotion targets for Node 0: null May 15 14:55:59.958494 kernel: Key type .fscrypt registered May 15 14:55:59.958507 kernel: Key type fscrypt-provisioning registered May 15 14:55:59.958556 kernel: ima: No TPM chip found, activating TPM-bypass! May 15 14:55:59.958575 kernel: ima: Allocated hash algorithm: sha1 May 15 14:55:59.958595 kernel: ima: No architecture policies found May 15 14:55:59.958610 kernel: clk: Disabling unused clocks May 15 14:55:59.958625 kernel: Warning: unable to open an initial console. May 15 14:55:59.958641 kernel: Freeing unused kernel image (initmem) memory: 54416K May 15 14:55:59.958656 kernel: Write protecting the kernel read-only data: 24576k May 15 14:55:59.958671 kernel: Freeing unused kernel image (rodata/data gap) memory: 296K May 15 14:55:59.958687 kernel: Run /init as init process May 15 14:55:59.958702 kernel: with arguments: May 15 14:55:59.958719 kernel: /init May 15 14:55:59.958734 kernel: with environment: May 15 14:55:59.958754 kernel: HOME=/ May 15 14:55:59.958769 kernel: TERM=linux May 15 14:55:59.958784 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 15 14:55:59.958801 systemd[1]: Successfully made /usr/ read-only. May 15 14:55:59.958822 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 15 14:55:59.958838 systemd[1]: Detected virtualization kvm. May 15 14:55:59.958854 systemd[1]: Detected architecture x86-64. May 15 14:55:59.958873 systemd[1]: Running in initrd. May 15 14:55:59.958888 systemd[1]: No hostname configured, using default hostname. May 15 14:55:59.958903 systemd[1]: Hostname set to . May 15 14:55:59.958918 systemd[1]: Initializing machine ID from VM UUID. May 15 14:55:59.958935 systemd[1]: Queued start job for default target initrd.target. May 15 14:55:59.958951 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 14:55:59.960923 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 14:55:59.960943 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 15 14:55:59.960986 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 15 14:55:59.963062 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 15 14:55:59.963094 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 15 14:55:59.963113 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 15 14:55:59.963136 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 15 14:55:59.963155 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 14:55:59.963171 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 15 14:55:59.963187 systemd[1]: Reached target paths.target - Path Units. May 15 14:55:59.963203 systemd[1]: Reached target slices.target - Slice Units. May 15 14:55:59.963220 systemd[1]: Reached target swap.target - Swaps. May 15 14:55:59.963237 systemd[1]: Reached target timers.target - Timer Units. May 15 14:55:59.963252 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 15 14:55:59.963273 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 15 14:55:59.963289 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 15 14:55:59.963305 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 15 14:55:59.963323 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 15 14:55:59.963339 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 15 14:55:59.963356 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 15 14:55:59.963371 systemd[1]: Reached target sockets.target - Socket Units. May 15 14:55:59.963389 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 15 14:55:59.963406 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 15 14:55:59.963428 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 15 14:55:59.963445 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 15 14:55:59.963460 systemd[1]: Starting systemd-fsck-usr.service... May 15 14:55:59.963477 systemd[1]: Starting systemd-journald.service - Journal Service... May 15 14:55:59.963511 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 15 14:55:59.963528 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 14:55:59.963544 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 15 14:55:59.963567 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 15 14:55:59.963645 systemd-journald[211]: Collecting audit messages is disabled. May 15 14:55:59.963689 systemd[1]: Finished systemd-fsck-usr.service. May 15 14:55:59.963706 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 15 14:55:59.963724 systemd-journald[211]: Journal started May 15 14:55:59.963759 systemd-journald[211]: Runtime Journal (/run/log/journal/15e48421cef04733a8fbe8f2d89510b4) is 4.9M, max 39.5M, 34.6M free. May 15 14:55:59.947313 systemd-modules-load[213]: Inserted module 'overlay' May 15 14:55:59.969038 systemd[1]: Started systemd-journald.service - Journal Service. May 15 14:55:59.982998 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 15 14:55:59.985097 systemd-modules-load[213]: Inserted module 'br_netfilter' May 15 14:56:00.021446 kernel: Bridge firewalling registered May 15 14:55:59.988211 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 15 14:56:00.029252 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 15 14:56:00.030319 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:00.031591 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 14:56:00.036811 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 15 14:56:00.039708 systemd-tmpfiles[227]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 15 14:56:00.040511 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 15 14:56:00.047199 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 15 14:56:00.049761 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 14:56:00.070433 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 14:56:00.072408 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 15 14:56:00.076952 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 15 14:56:00.085007 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 15 14:56:00.088203 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 15 14:56:00.118116 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=48287e633374b880fa618bd42bee102ae77c50831859c6cedd6ca9e1aec3dd5c May 15 14:56:00.140249 systemd-resolved[248]: Positive Trust Anchors: May 15 14:56:00.140265 systemd-resolved[248]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 15 14:56:00.140320 systemd-resolved[248]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 15 14:56:00.147878 systemd-resolved[248]: Defaulting to hostname 'linux'. May 15 14:56:00.151110 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 15 14:56:00.151804 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 15 14:56:00.252015 kernel: SCSI subsystem initialized May 15 14:56:00.262041 kernel: Loading iSCSI transport class v2.0-870. May 15 14:56:00.275017 kernel: iscsi: registered transport (tcp) May 15 14:56:00.299027 kernel: iscsi: registered transport (qla4xxx) May 15 14:56:00.299128 kernel: QLogic iSCSI HBA Driver May 15 14:56:00.328416 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 15 14:56:00.356589 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 15 14:56:00.359309 systemd[1]: Reached target network-pre.target - Preparation for Network. May 15 14:56:00.429884 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 15 14:56:00.433354 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 15 14:56:00.503023 kernel: raid6: avx2x4 gen() 11834 MB/s May 15 14:56:00.520031 kernel: raid6: avx2x2 gen() 12993 MB/s May 15 14:56:00.537158 kernel: raid6: avx2x1 gen() 10020 MB/s May 15 14:56:00.537280 kernel: raid6: using algorithm avx2x2 gen() 12993 MB/s May 15 14:56:00.555232 kernel: raid6: .... xor() 11409 MB/s, rmw enabled May 15 14:56:00.555349 kernel: raid6: using avx2x2 recovery algorithm May 15 14:56:00.587063 kernel: xor: automatically using best checksumming function avx May 15 14:56:00.852020 kernel: Btrfs loaded, zoned=no, fsverity=no May 15 14:56:00.863209 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 15 14:56:00.866905 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 14:56:00.904361 systemd-udevd[462]: Using default interface naming scheme 'v255'. May 15 14:56:00.914581 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 14:56:00.919224 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 15 14:56:00.954495 dracut-pre-trigger[467]: rd.md=0: removing MD RAID activation May 15 14:56:00.991762 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 15 14:56:00.994502 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 15 14:56:01.073648 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 15 14:56:01.078036 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 15 14:56:01.176835 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues May 15 14:56:01.201433 kernel: virtio_scsi virtio3: 2/0/0 default/read/poll queues May 15 14:56:01.228429 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) May 15 14:56:01.228642 kernel: ACPI: bus type USB registered May 15 14:56:01.228665 kernel: usbcore: registered new interface driver usbfs May 15 14:56:01.228684 kernel: usbcore: registered new interface driver hub May 15 14:56:01.228705 kernel: usbcore: registered new device driver usb May 15 14:56:01.228739 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 15 14:56:01.228760 kernel: GPT:9289727 != 125829119 May 15 14:56:01.228779 kernel: GPT:Alternate GPT header not at the end of the disk. May 15 14:56:01.228797 kernel: GPT:9289727 != 125829119 May 15 14:56:01.228816 kernel: GPT: Use GNU Parted to correct GPT errors. May 15 14:56:01.228836 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 14:56:01.228854 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues May 15 14:56:01.229084 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) May 15 14:56:01.229251 kernel: scsi host0: Virtio SCSI HBA May 15 14:56:01.249632 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller May 15 14:56:01.254799 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 May 15 14:56:01.255083 kernel: uhci_hcd 0000:00:01.2: detected 2 ports May 15 14:56:01.255286 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 May 15 14:56:01.255806 kernel: hub 1-0:1.0: USB hub found May 15 14:56:01.256045 kernel: hub 1-0:1.0: 2 ports detected May 15 14:56:01.272998 kernel: cryptd: max_cpu_qlen set to 1000 May 15 14:56:01.287011 kernel: libata version 3.00 loaded. May 15 14:56:01.291242 kernel: ata_piix 0000:00:01.1: version 2.13 May 15 14:56:01.305195 kernel: scsi host1: ata_piix May 15 14:56:01.305444 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 May 15 14:56:01.305471 kernel: scsi host2: ata_piix May 15 14:56:01.305676 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 lpm-pol 0 May 15 14:56:01.305710 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 lpm-pol 0 May 15 14:56:01.320665 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 14:56:01.321898 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:01.324923 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 15 14:56:01.328070 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 14:56:01.330586 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 15 14:56:01.410227 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 15 14:56:01.438026 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 15 14:56:01.439675 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:01.459267 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 15 14:56:01.472661 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 15 14:56:01.473228 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 15 14:56:01.476502 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 15 14:56:01.482996 kernel: AES CTR mode by8 optimization enabled May 15 14:56:01.521766 disk-uuid[601]: Primary Header is updated. May 15 14:56:01.521766 disk-uuid[601]: Secondary Entries is updated. May 15 14:56:01.521766 disk-uuid[601]: Secondary Header is updated. May 15 14:56:01.529095 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 15 14:56:01.531929 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 15 14:56:01.533370 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 14:56:01.534653 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 15 14:56:01.537259 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 15 14:56:01.539977 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 14:56:01.563676 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 14:56:01.563173 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 15 14:56:02.551995 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 15 14:56:02.553343 disk-uuid[610]: The operation has completed successfully. May 15 14:56:02.613624 systemd[1]: disk-uuid.service: Deactivated successfully. May 15 14:56:02.613847 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 15 14:56:02.669824 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 15 14:56:02.696991 sh[634]: Success May 15 14:56:02.720626 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 15 14:56:02.720736 kernel: device-mapper: uevent: version 1.0.3 May 15 14:56:02.721805 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev May 15 14:56:02.737021 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" May 15 14:56:02.815475 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 15 14:56:02.817296 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 15 14:56:02.830459 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 15 14:56:02.846006 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' May 15 14:56:02.848001 kernel: BTRFS: device fsid 2d504097-db49-4d66-a0d5-eeb665b21004 devid 1 transid 41 /dev/mapper/usr (253:0) scanned by mount (646) May 15 14:56:02.850604 kernel: BTRFS info (device dm-0): first mount of filesystem 2d504097-db49-4d66-a0d5-eeb665b21004 May 15 14:56:02.850720 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 15 14:56:02.850744 kernel: BTRFS info (device dm-0): using free-space-tree May 15 14:56:02.863208 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 15 14:56:02.864144 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 15 14:56:02.865043 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 15 14:56:02.866266 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 15 14:56:02.869875 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 15 14:56:02.901024 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (675) May 15 14:56:02.904358 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 14:56:02.904459 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 14:56:02.906003 kernel: BTRFS info (device vda6): using free-space-tree May 15 14:56:02.918055 kernel: BTRFS info (device vda6): last unmount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 14:56:02.918906 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 15 14:56:02.921684 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 15 14:56:03.069473 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 15 14:56:03.078161 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 15 14:56:03.143031 systemd-networkd[816]: lo: Link UP May 15 14:56:03.143045 systemd-networkd[816]: lo: Gained carrier May 15 14:56:03.148096 systemd-networkd[816]: Enumeration completed May 15 14:56:03.148916 systemd[1]: Started systemd-networkd.service - Network Configuration. May 15 14:56:03.150087 systemd-networkd[816]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 15 14:56:03.150157 systemd-networkd[816]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. May 15 14:56:03.151467 systemd-networkd[816]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 15 14:56:03.151473 systemd-networkd[816]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. May 15 14:56:03.152389 systemd-networkd[816]: eth0: Link UP May 15 14:56:03.152400 systemd-networkd[816]: eth0: Gained carrier May 15 14:56:03.152419 systemd-networkd[816]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 15 14:56:03.154294 systemd[1]: Reached target network.target - Network. May 15 14:56:03.157260 systemd-networkd[816]: eth1: Link UP May 15 14:56:03.157266 systemd-networkd[816]: eth1: Gained carrier May 15 14:56:03.157286 systemd-networkd[816]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 15 14:56:03.168888 ignition[724]: Ignition 2.21.0 May 15 14:56:03.168910 ignition[724]: Stage: fetch-offline May 15 14:56:03.169014 ignition[724]: no configs at "/usr/lib/ignition/base.d" May 15 14:56:03.169032 ignition[724]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:03.172256 ignition[724]: parsed url from cmdline: "" May 15 14:56:03.172283 ignition[724]: no config URL provided May 15 14:56:03.172304 ignition[724]: reading system config file "/usr/lib/ignition/user.ign" May 15 14:56:03.173225 systemd-networkd[816]: eth0: DHCPv4 address 24.199.109.202/20, gateway 24.199.96.1 acquired from 169.254.169.253 May 15 14:56:03.172336 ignition[724]: no config at "/usr/lib/ignition/user.ign" May 15 14:56:03.177467 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 15 14:56:03.172350 ignition[724]: failed to fetch config: resource requires networking May 15 14:56:03.173095 ignition[724]: Ignition finished successfully May 15 14:56:03.179184 systemd-networkd[816]: eth1: DHCPv4 address 10.124.0.28/20 acquired from 169.254.169.253 May 15 14:56:03.184388 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 15 14:56:03.227171 ignition[825]: Ignition 2.21.0 May 15 14:56:03.228095 ignition[825]: Stage: fetch May 15 14:56:03.228451 ignition[825]: no configs at "/usr/lib/ignition/base.d" May 15 14:56:03.228473 ignition[825]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:03.228630 ignition[825]: parsed url from cmdline: "" May 15 14:56:03.228636 ignition[825]: no config URL provided May 15 14:56:03.228646 ignition[825]: reading system config file "/usr/lib/ignition/user.ign" May 15 14:56:03.228660 ignition[825]: no config at "/usr/lib/ignition/user.ign" May 15 14:56:03.228711 ignition[825]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 May 15 14:56:03.246709 ignition[825]: GET result: OK May 15 14:56:03.246987 ignition[825]: parsing config with SHA512: 0d8f4760086f8df2e18d751a2444f453bf813d28801df795c507a5c04d78d7b6b11c9075189b2c7b8d7e79677ddc8424ae8903f8df57a628b2cb87364d52dab9 May 15 14:56:03.255034 unknown[825]: fetched base config from "system" May 15 14:56:03.255670 ignition[825]: fetch: fetch complete May 15 14:56:03.255046 unknown[825]: fetched base config from "system" May 15 14:56:03.255678 ignition[825]: fetch: fetch passed May 15 14:56:03.255053 unknown[825]: fetched user config from "digitalocean" May 15 14:56:03.255769 ignition[825]: Ignition finished successfully May 15 14:56:03.261351 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 15 14:56:03.269190 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 15 14:56:03.313453 ignition[831]: Ignition 2.21.0 May 15 14:56:03.313476 ignition[831]: Stage: kargs May 15 14:56:03.313807 ignition[831]: no configs at "/usr/lib/ignition/base.d" May 15 14:56:03.313823 ignition[831]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:03.322047 ignition[831]: kargs: kargs passed May 15 14:56:03.322240 ignition[831]: Ignition finished successfully May 15 14:56:03.326523 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 15 14:56:03.329703 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 15 14:56:03.370922 ignition[837]: Ignition 2.21.0 May 15 14:56:03.371749 ignition[837]: Stage: disks May 15 14:56:03.372022 ignition[837]: no configs at "/usr/lib/ignition/base.d" May 15 14:56:03.372033 ignition[837]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:03.374313 ignition[837]: disks: disks passed May 15 14:56:03.374487 ignition[837]: Ignition finished successfully May 15 14:56:03.377567 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 15 14:56:03.378744 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 15 14:56:03.379185 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 15 14:56:03.380164 systemd[1]: Reached target local-fs.target - Local File Systems. May 15 14:56:03.381028 systemd[1]: Reached target sysinit.target - System Initialization. May 15 14:56:03.381864 systemd[1]: Reached target basic.target - Basic System. May 15 14:56:03.384290 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 15 14:56:03.420092 systemd-fsck[846]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 15 14:56:03.424269 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 15 14:56:03.428460 systemd[1]: Mounting sysroot.mount - /sysroot... May 15 14:56:03.587996 kernel: EXT4-fs (vda9): mounted filesystem f7dea4bd-2644-4592-b85b-330f322c4d2b r/w with ordered data mode. Quota mode: none. May 15 14:56:03.589621 systemd[1]: Mounted sysroot.mount - /sysroot. May 15 14:56:03.591680 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 15 14:56:03.596563 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 15 14:56:03.599220 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 15 14:56:03.607203 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... May 15 14:56:03.612157 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... May 15 14:56:03.623148 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 15 14:56:03.624408 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 15 14:56:03.628847 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 15 14:56:03.631886 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (854) May 15 14:56:03.633393 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 14:56:03.633456 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 14:56:03.636120 kernel: BTRFS info (device vda6): using free-space-tree May 15 14:56:03.644247 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 15 14:56:03.675428 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 15 14:56:03.736116 coreos-metadata[857]: May 15 14:56:03.736 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 14:56:03.745008 coreos-metadata[857]: May 15 14:56:03.744 INFO Fetch successful May 15 14:56:03.751091 coreos-metadata[856]: May 15 14:56:03.751 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 14:56:03.753735 coreos-metadata[857]: May 15 14:56:03.753 INFO wrote hostname ci-4334.0.0-a-c1acadba3f to /sysroot/etc/hostname May 15 14:56:03.758183 initrd-setup-root[885]: cut: /sysroot/etc/passwd: No such file or directory May 15 14:56:03.758471 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 15 14:56:03.762309 coreos-metadata[856]: May 15 14:56:03.762 INFO Fetch successful May 15 14:56:03.770929 initrd-setup-root[892]: cut: /sysroot/etc/group: No such file or directory May 15 14:56:03.771291 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. May 15 14:56:03.771469 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. May 15 14:56:03.779681 initrd-setup-root[900]: cut: /sysroot/etc/shadow: No such file or directory May 15 14:56:03.786251 initrd-setup-root[907]: cut: /sysroot/etc/gshadow: No such file or directory May 15 14:56:03.951718 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 15 14:56:03.954249 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 15 14:56:03.955794 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 15 14:56:03.984925 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 15 14:56:03.986060 kernel: BTRFS info (device vda6): last unmount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 14:56:04.010012 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 15 14:56:04.030354 ignition[976]: INFO : Ignition 2.21.0 May 15 14:56:04.030354 ignition[976]: INFO : Stage: mount May 15 14:56:04.033697 ignition[976]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 14:56:04.033697 ignition[976]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:04.033697 ignition[976]: INFO : mount: mount passed May 15 14:56:04.033697 ignition[976]: INFO : Ignition finished successfully May 15 14:56:04.035795 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 15 14:56:04.038907 systemd[1]: Starting ignition-files.service - Ignition (files)... May 15 14:56:04.065694 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 15 14:56:04.095015 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (988) May 15 14:56:04.097191 kernel: BTRFS info (device vda6): first mount of filesystem afd0c70c-d15e-448c-8325-f96e3c3ed3a5 May 15 14:56:04.097301 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 15 14:56:04.098031 kernel: BTRFS info (device vda6): using free-space-tree May 15 14:56:04.106643 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 15 14:56:04.150821 ignition[1005]: INFO : Ignition 2.21.0 May 15 14:56:04.150821 ignition[1005]: INFO : Stage: files May 15 14:56:04.153569 ignition[1005]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 14:56:04.153569 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:04.158053 ignition[1005]: DEBUG : files: compiled without relabeling support, skipping May 15 14:56:04.160710 ignition[1005]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 15 14:56:04.160710 ignition[1005]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 15 14:56:04.166117 ignition[1005]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 15 14:56:04.167195 ignition[1005]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 15 14:56:04.168499 unknown[1005]: wrote ssh authorized keys file for user: core May 15 14:56:04.169987 ignition[1005]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 15 14:56:04.173265 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 15 14:56:04.174508 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 May 15 14:56:04.230901 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 15 14:56:04.452008 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 15 14:56:04.452008 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 15 14:56:04.452008 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 15 14:56:04.452008 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 14:56:04.458165 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 May 15 14:56:04.895249 systemd-networkd[816]: eth0: Gained IPv6LL May 15 14:56:05.008020 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK May 15 14:56:05.151229 systemd-networkd[816]: eth1: Gained IPv6LL May 15 14:56:05.985596 ignition[1005]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 15 14:56:05.985596 ignition[1005]: INFO : files: op(b): [started] processing unit "prepare-helm.service" May 15 14:56:05.988323 ignition[1005]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 15 14:56:05.989710 ignition[1005]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 15 14:56:05.989710 ignition[1005]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" May 15 14:56:05.989710 ignition[1005]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" May 15 14:56:05.991672 ignition[1005]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" May 15 14:56:05.991672 ignition[1005]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" May 15 14:56:05.991672 ignition[1005]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" May 15 14:56:05.991672 ignition[1005]: INFO : files: files passed May 15 14:56:05.991672 ignition[1005]: INFO : Ignition finished successfully May 15 14:56:05.992796 systemd[1]: Finished ignition-files.service - Ignition (files). May 15 14:56:05.996199 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 15 14:56:05.998925 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 15 14:56:06.013361 systemd[1]: ignition-quench.service: Deactivated successfully. May 15 14:56:06.013607 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 15 14:56:06.028275 initrd-setup-root-after-ignition[1035]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 15 14:56:06.028275 initrd-setup-root-after-ignition[1035]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 15 14:56:06.030619 initrd-setup-root-after-ignition[1039]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 15 14:56:06.032197 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 15 14:56:06.033649 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 15 14:56:06.035758 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 15 14:56:06.098854 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 15 14:56:06.099024 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 15 14:56:06.100237 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 15 14:56:06.100886 systemd[1]: Reached target initrd.target - Initrd Default Target. May 15 14:56:06.101844 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 15 14:56:06.102927 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 15 14:56:06.138038 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 15 14:56:06.141341 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 15 14:56:06.169881 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 15 14:56:06.170706 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 14:56:06.171764 systemd[1]: Stopped target timers.target - Timer Units. May 15 14:56:06.172724 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 15 14:56:06.172952 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 15 14:56:06.173982 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 15 14:56:06.174837 systemd[1]: Stopped target basic.target - Basic System. May 15 14:56:06.175724 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 15 14:56:06.176443 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 15 14:56:06.177100 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 15 14:56:06.177807 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 15 14:56:06.178599 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 15 14:56:06.179164 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 15 14:56:06.180238 systemd[1]: Stopped target sysinit.target - System Initialization. May 15 14:56:06.180920 systemd[1]: Stopped target local-fs.target - Local File Systems. May 15 14:56:06.181578 systemd[1]: Stopped target swap.target - Swaps. May 15 14:56:06.182158 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 15 14:56:06.182314 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 15 14:56:06.183363 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 15 14:56:06.184491 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 14:56:06.185079 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 15 14:56:06.185229 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 14:56:06.185834 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 15 14:56:06.186113 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 15 14:56:06.187862 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 15 14:56:06.188200 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 15 14:56:06.189320 systemd[1]: ignition-files.service: Deactivated successfully. May 15 14:56:06.189576 systemd[1]: Stopped ignition-files.service - Ignition (files). May 15 14:56:06.190388 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. May 15 14:56:06.190541 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 15 14:56:06.192868 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 15 14:56:06.196367 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 15 14:56:06.196887 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 15 14:56:06.197247 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 15 14:56:06.200217 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 15 14:56:06.200376 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 15 14:56:06.209908 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 15 14:56:06.212200 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 15 14:56:06.234597 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 15 14:56:06.243085 ignition[1059]: INFO : Ignition 2.21.0 May 15 14:56:06.243085 ignition[1059]: INFO : Stage: umount May 15 14:56:06.268639 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" May 15 14:56:06.268639 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 15 14:56:06.268639 ignition[1059]: INFO : umount: umount passed May 15 14:56:06.268639 ignition[1059]: INFO : Ignition finished successfully May 15 14:56:06.266354 systemd[1]: ignition-mount.service: Deactivated successfully. May 15 14:56:06.266510 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 15 14:56:06.285741 systemd[1]: ignition-disks.service: Deactivated successfully. May 15 14:56:06.286353 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 15 14:56:06.289724 systemd[1]: ignition-kargs.service: Deactivated successfully. May 15 14:56:06.289847 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 15 14:56:06.290489 systemd[1]: ignition-fetch.service: Deactivated successfully. May 15 14:56:06.290577 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 15 14:56:06.291850 systemd[1]: Stopped target network.target - Network. May 15 14:56:06.294393 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 15 14:56:06.294534 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 15 14:56:06.295242 systemd[1]: Stopped target paths.target - Path Units. May 15 14:56:06.298244 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 15 14:56:06.302154 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 14:56:06.303310 systemd[1]: Stopped target slices.target - Slice Units. May 15 14:56:06.304002 systemd[1]: Stopped target sockets.target - Socket Units. May 15 14:56:06.304876 systemd[1]: iscsid.socket: Deactivated successfully. May 15 14:56:06.304980 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 15 14:56:06.305644 systemd[1]: iscsiuio.socket: Deactivated successfully. May 15 14:56:06.305703 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 15 14:56:06.306413 systemd[1]: ignition-setup.service: Deactivated successfully. May 15 14:56:06.306513 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 15 14:56:06.307203 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 15 14:56:06.307278 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 15 14:56:06.308198 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 15 14:56:06.308854 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 15 14:56:06.310650 systemd[1]: sysroot-boot.service: Deactivated successfully. May 15 14:56:06.310811 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 15 14:56:06.312326 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 15 14:56:06.312501 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 15 14:56:06.318535 systemd[1]: systemd-networkd.service: Deactivated successfully. May 15 14:56:06.318738 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 15 14:56:06.321803 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 15 14:56:06.322264 systemd[1]: systemd-resolved.service: Deactivated successfully. May 15 14:56:06.322423 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 15 14:56:06.327347 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 15 14:56:06.328989 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 15 14:56:06.329677 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 15 14:56:06.329751 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 15 14:56:06.332248 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 15 14:56:06.332725 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 15 14:56:06.332821 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 15 14:56:06.333439 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 15 14:56:06.333511 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 15 14:56:06.336226 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 15 14:56:06.336319 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 15 14:56:06.336984 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 15 14:56:06.337071 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 14:56:06.341199 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 14:56:06.346508 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 15 14:56:06.346631 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 15 14:56:06.355699 systemd[1]: systemd-udevd.service: Deactivated successfully. May 15 14:56:06.356022 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 14:56:06.357532 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 15 14:56:06.357667 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 15 14:56:06.358694 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 15 14:56:06.358750 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 15 14:56:06.361339 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 15 14:56:06.361448 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 15 14:56:06.364714 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 15 14:56:06.364810 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 15 14:56:06.365232 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 15 14:56:06.365298 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 15 14:56:06.374311 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 15 14:56:06.376196 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 15 14:56:06.376952 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 15 14:56:06.378189 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 15 14:56:06.378300 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 14:56:06.379906 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 15 14:56:06.380443 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 14:56:06.381907 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 15 14:56:06.382064 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 15 14:56:06.383501 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 14:56:06.383573 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:06.387141 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. May 15 14:56:06.388060 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. May 15 14:56:06.388127 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 15 14:56:06.388181 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 15 14:56:06.391204 systemd[1]: network-cleanup.service: Deactivated successfully. May 15 14:56:06.391465 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 15 14:56:06.394882 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 15 14:56:06.395097 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 15 14:56:06.399636 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 15 14:56:06.401752 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 15 14:56:06.438631 systemd[1]: Switching root. May 15 14:56:06.511927 systemd-journald[211]: Journal stopped May 15 14:56:08.069826 systemd-journald[211]: Received SIGTERM from PID 1 (systemd). May 15 14:56:08.070006 kernel: SELinux: policy capability network_peer_controls=1 May 15 14:56:08.070054 kernel: SELinux: policy capability open_perms=1 May 15 14:56:08.070074 kernel: SELinux: policy capability extended_socket_class=1 May 15 14:56:08.070103 kernel: SELinux: policy capability always_check_network=0 May 15 14:56:08.070145 kernel: SELinux: policy capability cgroup_seclabel=1 May 15 14:56:08.070165 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 15 14:56:08.070183 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 15 14:56:08.070202 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 15 14:56:08.070232 kernel: SELinux: policy capability userspace_initial_context=0 May 15 14:56:08.070250 kernel: audit: type=1403 audit(1747320966.687:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 15 14:56:08.070286 systemd[1]: Successfully loaded SELinux policy in 58.299ms. May 15 14:56:08.070330 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 42.821ms. May 15 14:56:08.070346 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 15 14:56:08.070359 systemd[1]: Detected virtualization kvm. May 15 14:56:08.070375 systemd[1]: Detected architecture x86-64. May 15 14:56:08.070389 systemd[1]: Detected first boot. May 15 14:56:08.070409 systemd[1]: Hostname set to . May 15 14:56:08.070428 systemd[1]: Initializing machine ID from VM UUID. May 15 14:56:08.070440 kernel: Guest personality initialized and is inactive May 15 14:56:08.070452 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 15 14:56:08.070470 kernel: Initialized host personality May 15 14:56:08.070481 kernel: NET: Registered PF_VSOCK protocol family May 15 14:56:08.070493 zram_generator::config[1102]: No configuration found. May 15 14:56:08.070507 systemd[1]: Populated /etc with preset unit settings. May 15 14:56:08.070524 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 15 14:56:08.070549 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 15 14:56:08.070563 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 15 14:56:08.070577 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 15 14:56:08.070592 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 15 14:56:08.070604 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 15 14:56:08.070616 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 15 14:56:08.070628 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 15 14:56:08.070640 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 15 14:56:08.070658 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 15 14:56:08.070677 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 15 14:56:08.070694 systemd[1]: Created slice user.slice - User and Session Slice. May 15 14:56:08.070706 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 15 14:56:08.070718 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 15 14:56:08.070735 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 15 14:56:08.070750 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 15 14:56:08.070769 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 15 14:56:08.070781 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 15 14:56:08.070793 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 15 14:56:08.070808 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 15 14:56:08.070822 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 15 14:56:08.070834 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 15 14:56:08.070851 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 15 14:56:08.070863 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 15 14:56:08.070881 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 15 14:56:08.070894 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 15 14:56:08.070910 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 15 14:56:08.070925 systemd[1]: Reached target slices.target - Slice Units. May 15 14:56:08.070941 systemd[1]: Reached target swap.target - Swaps. May 15 14:56:08.070975 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 15 14:56:08.070991 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 15 14:56:08.071011 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 15 14:56:08.071024 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 15 14:56:08.071049 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 15 14:56:08.071070 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 15 14:56:08.071082 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 15 14:56:08.071095 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 15 14:56:08.071107 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 15 14:56:08.071121 systemd[1]: Mounting media.mount - External Media Directory... May 15 14:56:08.071133 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:08.071146 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 15 14:56:08.071161 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 15 14:56:08.071178 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 15 14:56:08.071191 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 15 14:56:08.071206 systemd[1]: Reached target machines.target - Containers. May 15 14:56:08.071225 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 15 14:56:08.071238 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 14:56:08.071255 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 15 14:56:08.071272 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 15 14:56:08.071288 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 14:56:08.071305 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 15 14:56:08.071332 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 14:56:08.071355 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 15 14:56:08.071376 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 14:56:08.071395 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 15 14:56:08.071415 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 15 14:56:08.071430 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 15 14:56:08.071443 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 15 14:56:08.071455 systemd[1]: Stopped systemd-fsck-usr.service. May 15 14:56:08.071491 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 14:56:08.071510 systemd[1]: Starting systemd-journald.service - Journal Service... May 15 14:56:08.071530 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 15 14:56:08.071548 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 15 14:56:08.071561 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 15 14:56:08.071573 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 15 14:56:08.071594 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 15 14:56:08.071632 systemd[1]: verity-setup.service: Deactivated successfully. May 15 14:56:08.071653 systemd[1]: Stopped verity-setup.service. May 15 14:56:08.071670 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:08.071697 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 15 14:56:08.071716 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 15 14:56:08.071733 systemd[1]: Mounted media.mount - External Media Directory. May 15 14:56:08.071756 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 15 14:56:08.071773 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 15 14:56:08.071789 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 15 14:56:08.071807 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 15 14:56:08.071826 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 15 14:56:08.071844 kernel: fuse: init (API version 7.41) May 15 14:56:08.071878 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 15 14:56:08.071898 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 14:56:08.071911 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 14:56:08.071923 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 14:56:08.071935 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 14:56:08.071947 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 15 14:56:08.071977 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 15 14:56:08.071990 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 15 14:56:08.072006 systemd[1]: Reached target network-pre.target - Preparation for Network. May 15 14:56:08.072039 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 15 14:56:08.072054 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 15 14:56:08.072067 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 15 14:56:08.072079 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 15 14:56:08.072096 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 15 14:56:08.072111 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 15 14:56:08.072132 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 15 14:56:08.072144 systemd[1]: Reached target local-fs.target - Local File Systems. May 15 14:56:08.072157 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 15 14:56:08.072179 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 15 14:56:08.072193 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 14:56:08.072205 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 15 14:56:08.072221 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 14:56:08.072237 kernel: loop: module loaded May 15 14:56:08.072299 systemd-journald[1172]: Collecting audit messages is disabled. May 15 14:56:08.072343 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 15 14:56:08.072363 systemd-journald[1172]: Journal started May 15 14:56:08.072391 systemd-journald[1172]: Runtime Journal (/run/log/journal/15e48421cef04733a8fbe8f2d89510b4) is 4.9M, max 39.5M, 34.6M free. May 15 14:56:07.679657 systemd[1]: Queued start job for default target multi-user.target. May 15 14:56:07.695691 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 15 14:56:07.696449 systemd[1]: systemd-journald.service: Deactivated successfully. May 15 14:56:08.087985 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 15 14:56:08.088071 systemd[1]: Started systemd-journald.service - Journal Service. May 15 14:56:08.079460 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 14:56:08.079711 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 14:56:08.080511 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 15 14:56:08.115720 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 15 14:56:08.117100 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 14:56:08.129011 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 15 14:56:08.143127 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. May 15 14:56:08.143852 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. May 15 14:56:08.164450 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 15 14:56:08.176735 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 15 14:56:08.201084 kernel: loop0: detected capacity change from 0 to 146240 May 15 14:56:08.205248 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 15 14:56:08.206409 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 15 14:56:08.216224 systemd-journald[1172]: Time spent on flushing to /var/log/journal/15e48421cef04733a8fbe8f2d89510b4 is 100.669ms for 1014 entries. May 15 14:56:08.216224 systemd-journald[1172]: System Journal (/var/log/journal/15e48421cef04733a8fbe8f2d89510b4) is 8M, max 195.6M, 187.6M free. May 15 14:56:08.373489 systemd-journald[1172]: Received client request to flush runtime journal. May 15 14:56:08.373543 kernel: ACPI: bus type drm_connector registered May 15 14:56:08.373567 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 15 14:56:08.373587 kernel: loop1: detected capacity change from 0 to 8 May 15 14:56:08.373607 kernel: loop2: detected capacity change from 0 to 113872 May 15 14:56:08.215360 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 15 14:56:08.268610 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 15 14:56:08.276235 systemd[1]: modprobe@drm.service: Deactivated successfully. May 15 14:56:08.276498 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 15 14:56:08.277695 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 15 14:56:08.311261 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 15 14:56:08.344138 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 15 14:56:08.375668 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 15 14:56:08.383060 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 15 14:56:08.409117 kernel: loop3: detected capacity change from 0 to 218376 May 15 14:56:08.444473 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 15 14:56:08.446121 kernel: loop4: detected capacity change from 0 to 146240 May 15 14:56:08.449599 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 15 14:56:08.470993 kernel: loop5: detected capacity change from 0 to 8 May 15 14:56:08.494995 kernel: loop6: detected capacity change from 0 to 113872 May 15 14:56:08.536088 kernel: loop7: detected capacity change from 0 to 218376 May 15 14:56:08.560030 (sd-merge)[1251]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. May 15 14:56:08.560646 (sd-merge)[1251]: Merged extensions into '/usr'. May 15 14:56:08.578509 systemd[1]: Reload requested from client PID 1196 ('systemd-sysext') (unit systemd-sysext.service)... May 15 14:56:08.578533 systemd[1]: Reloading... May 15 14:56:08.605678 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. May 15 14:56:08.613145 systemd-tmpfiles[1253]: ACLs are not supported, ignoring. May 15 14:56:08.772006 zram_generator::config[1281]: No configuration found. May 15 14:56:09.041334 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 14:56:09.161321 ldconfig[1191]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 15 14:56:09.174201 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 15 14:56:09.174558 systemd[1]: Reloading finished in 593 ms. May 15 14:56:09.193345 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 15 14:56:09.194266 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 15 14:56:09.195086 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 15 14:56:09.207175 systemd[1]: Starting ensure-sysext.service... May 15 14:56:09.213232 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 15 14:56:09.246214 systemd[1]: Reload requested from client PID 1325 ('systemctl') (unit ensure-sysext.service)... May 15 14:56:09.246242 systemd[1]: Reloading... May 15 14:56:09.307222 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 15 14:56:09.307629 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 15 14:56:09.307971 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 15 14:56:09.308456 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 15 14:56:09.312277 systemd-tmpfiles[1326]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 15 14:56:09.312950 systemd-tmpfiles[1326]: ACLs are not supported, ignoring. May 15 14:56:09.313070 systemd-tmpfiles[1326]: ACLs are not supported, ignoring. May 15 14:56:09.336271 systemd-tmpfiles[1326]: Detected autofs mount point /boot during canonicalization of boot. May 15 14:56:09.336290 systemd-tmpfiles[1326]: Skipping /boot May 15 14:56:09.361050 zram_generator::config[1350]: No configuration found. May 15 14:56:09.408289 systemd-tmpfiles[1326]: Detected autofs mount point /boot during canonicalization of boot. May 15 14:56:09.408309 systemd-tmpfiles[1326]: Skipping /boot May 15 14:56:09.549167 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 14:56:09.700133 systemd[1]: Reloading finished in 453 ms. May 15 14:56:09.711538 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 15 14:56:09.728966 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 15 14:56:09.737171 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 15 14:56:09.740302 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 15 14:56:09.747341 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 15 14:56:09.751031 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 15 14:56:09.757550 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 15 14:56:09.762640 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 15 14:56:09.771364 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.771600 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 14:56:09.775324 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 14:56:09.779365 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 14:56:09.789897 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 14:56:09.790733 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 14:56:09.790943 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 14:56:09.791071 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.796726 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.797051 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 14:56:09.797302 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 14:56:09.797401 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 14:56:09.800226 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 15 14:56:09.802080 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.808475 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.808726 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 14:56:09.815304 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 15 14:56:09.816330 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 14:56:09.816499 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 14:56:09.816659 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:09.823837 systemd[1]: Finished ensure-sysext.service. May 15 14:56:09.837343 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 15 14:56:09.860611 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 15 14:56:09.872652 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 15 14:56:09.874671 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 15 14:56:09.876458 systemd-udevd[1402]: Using default interface naming scheme 'v255'. May 15 14:56:09.909586 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 15 14:56:09.923303 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 15 14:56:09.928071 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 15 14:56:09.935575 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 15 14:56:09.938807 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 14:56:09.948049 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 14:56:09.953341 systemd[1]: modprobe@drm.service: Deactivated successfully. May 15 14:56:09.953585 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 15 14:56:09.954911 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 14:56:09.955881 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 14:56:09.957807 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 14:56:09.967894 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 14:56:09.969981 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 14:56:09.971363 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 14:56:09.998482 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 15 14:56:10.011407 augenrules[1457]: No rules May 15 14:56:10.014890 systemd[1]: audit-rules.service: Deactivated successfully. May 15 14:56:10.017133 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 15 14:56:10.033782 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 15 14:56:10.160721 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. May 15 14:56:10.160876 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 15 14:56:10.163996 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... May 15 14:56:10.165050 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:10.165219 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 15 14:56:10.166857 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 15 14:56:10.171297 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 15 14:56:10.179138 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 15 14:56:10.179761 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 15 14:56:10.179826 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 15 14:56:10.179860 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 15 14:56:10.179878 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 15 14:56:10.208001 kernel: ISO 9660 Extensions: RRIP_1991A May 15 14:56:10.219320 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. May 15 14:56:10.226075 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 15 14:56:10.226682 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 15 14:56:10.228584 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 15 14:56:10.230056 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 15 14:56:10.231223 systemd[1]: modprobe@loop.service: Deactivated successfully. May 15 14:56:10.231439 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 15 14:56:10.238260 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 15 14:56:10.238336 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 15 14:56:10.276437 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 15 14:56:10.278149 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 15 14:56:10.311632 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 15 14:56:10.396242 kernel: mousedev: PS/2 mouse device common for all mice May 15 14:56:10.480544 systemd-networkd[1431]: lo: Link UP May 15 14:56:10.481050 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 May 15 14:56:10.480559 systemd-networkd[1431]: lo: Gained carrier May 15 14:56:10.492672 systemd-networkd[1431]: Enumeration completed May 15 14:56:10.492852 systemd[1]: Started systemd-networkd.service - Network Configuration. May 15 14:56:10.496696 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 15 14:56:10.497459 systemd-networkd[1431]: eth0: Configuring with /run/systemd/network/10-52:79:ed:bc:3f:4e.network. May 15 14:56:10.500471 systemd-networkd[1431]: eth1: Configuring with /run/systemd/network/10-9e:a9:bb:9b:f0:e4.network. May 15 14:56:10.504195 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 15 14:56:10.504922 systemd-networkd[1431]: eth0: Link UP May 15 14:56:10.505126 systemd-networkd[1431]: eth0: Gained carrier May 15 14:56:10.510517 systemd-networkd[1431]: eth1: Link UP May 15 14:56:10.512902 systemd-networkd[1431]: eth1: Gained carrier May 15 14:56:10.520988 kernel: ACPI: button: Power Button [PWRF] May 15 14:56:10.537424 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 15 14:56:10.554402 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 May 15 14:56:10.556217 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 15 14:56:10.564685 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 15 14:56:10.565968 systemd[1]: Reached target time-set.target - System Time Set. May 15 14:56:10.578081 systemd-resolved[1401]: Positive Trust Anchors: May 15 14:56:10.578524 systemd-resolved[1401]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 15 14:56:10.578629 systemd-resolved[1401]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 15 14:56:10.579991 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 May 15 14:56:10.580074 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console May 15 14:56:10.587498 kernel: Console: switching to colour dummy device 80x25 May 15 14:56:10.586373 systemd-resolved[1401]: Using system hostname 'ci-4334.0.0-a-c1acadba3f'. May 15 14:56:10.588421 kernel: [drm] features: -virgl +edid -resource_blob -host_visible May 15 14:56:10.588478 kernel: [drm] features: -context_init May 15 14:56:10.591077 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 15 14:56:10.591270 systemd[1]: Reached target network.target - Network. May 15 14:56:10.591991 kernel: [drm] number of scanouts: 1 May 15 14:56:10.592080 kernel: [drm] number of cap sets: 0 May 15 14:56:10.592037 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 15 14:56:10.592139 systemd[1]: Reached target sysinit.target - System Initialization. May 15 14:56:10.592306 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 15 14:56:10.592403 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 15 14:56:10.592456 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 15 14:56:10.592651 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 15 14:56:10.592786 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 15 14:56:10.592845 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 15 14:56:10.592903 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 15 14:56:10.592930 systemd[1]: Reached target paths.target - Path Units. May 15 14:56:10.592983 systemd[1]: Reached target timers.target - Timer Units. May 15 14:56:10.599007 kernel: [drm] Initialized virtio_gpu 0.1.0 for 0000:00:02.0 on minor 0 May 15 14:56:10.599672 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 15 14:56:10.602602 systemd[1]: Starting docker.socket - Docker Socket for the API... May 15 14:56:10.607941 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 15 14:56:10.608439 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 15 14:56:10.608522 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 15 14:56:10.611641 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 15 14:56:10.612598 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 15 14:56:10.614470 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 15 14:56:10.615683 systemd[1]: Reached target sockets.target - Socket Units. May 15 14:56:10.615782 systemd[1]: Reached target basic.target - Basic System. May 15 14:56:10.615913 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 15 14:56:10.615944 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 15 14:56:10.619293 systemd[1]: Starting containerd.service - containerd container runtime... May 15 14:56:10.623878 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 15 14:56:10.627998 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 15 14:56:10.631382 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 15 14:56:10.633276 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 15 14:56:10.642171 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 15 14:56:10.642313 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 15 14:56:10.650890 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 15 14:56:10.657428 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 15 14:56:10.667200 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 15 14:56:10.673861 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 15 14:56:10.689381 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 15 14:56:10.711382 systemd[1]: Starting systemd-logind.service - User Login Management... May 15 14:56:10.714252 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 15 14:56:10.714994 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 15 14:56:10.720043 jq[1516]: false May 15 14:56:10.721214 systemd[1]: Starting update-engine.service - Update Engine... May 15 14:56:10.724944 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 15 14:56:10.744856 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 15 14:56:10.745286 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 15 14:56:10.745467 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 15 14:56:10.754996 coreos-metadata[1513]: May 15 14:56:10.751 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 14:56:10.767002 extend-filesystems[1517]: Found loop4 May 15 14:56:10.767002 extend-filesystems[1517]: Found loop5 May 15 14:56:10.767002 extend-filesystems[1517]: Found loop6 May 15 14:56:10.767002 extend-filesystems[1517]: Found loop7 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda May 15 14:56:10.767002 extend-filesystems[1517]: Found vda1 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda2 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda3 May 15 14:56:10.767002 extend-filesystems[1517]: Found usr May 15 14:56:10.767002 extend-filesystems[1517]: Found vda4 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda6 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda7 May 15 14:56:10.767002 extend-filesystems[1517]: Found vda9 May 15 14:56:10.767002 extend-filesystems[1517]: Checking size of /dev/vda9 May 15 14:56:10.759664 oslogin_cache_refresh[1518]: Refreshing passwd entry cache May 15 14:56:10.772523 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Refreshing passwd entry cache May 15 14:56:10.772734 jq[1532]: true May 15 14:56:10.789996 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Failure getting users, quitting May 15 14:56:10.789996 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 15 14:56:10.789996 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Refreshing group entry cache May 15 14:56:10.790205 coreos-metadata[1513]: May 15 14:56:10.784 INFO Fetch successful May 15 14:56:10.781493 oslogin_cache_refresh[1518]: Failure getting users, quitting May 15 14:56:10.781517 oslogin_cache_refresh[1518]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 15 14:56:10.781581 oslogin_cache_refresh[1518]: Refreshing group entry cache May 15 14:56:10.792804 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 15 14:56:11.389806 systemd-timesyncd[1416]: Contacted time server 45.79.189.79:123 (0.flatcar.pool.ntp.org). May 15 14:56:11.393767 oslogin_cache_refresh[1518]: Failure getting groups, quitting May 15 14:56:11.395685 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Failure getting groups, quitting May 15 14:56:11.395685 google_oslogin_nss_cache[1518]: oslogin_cache_refresh[1518]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 15 14:56:11.389884 systemd-timesyncd[1416]: Initial clock synchronization to Thu 2025-05-15 14:56:11.389653 UTC. May 15 14:56:11.393783 oslogin_cache_refresh[1518]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 15 14:56:11.390627 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 15 14:56:11.392350 systemd-resolved[1401]: Clock change detected. Flushing caches. May 15 14:56:11.402597 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 15 14:56:11.403505 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 15 14:56:11.430830 extend-filesystems[1517]: Resized partition /dev/vda9 May 15 14:56:11.437513 extend-filesystems[1560]: resize2fs 1.47.2 (1-Jan-2025) May 15 14:56:11.442384 dbus-daemon[1514]: [system] SELinux support is enabled May 15 14:56:11.448380 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 15 14:56:11.457337 systemd[1]: motdgen.service: Deactivated successfully. May 15 14:56:11.459654 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 15 14:56:11.466091 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 15 14:56:11.466139 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 15 14:56:11.466328 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 15 14:56:11.466426 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). May 15 14:56:11.466444 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 15 14:56:11.476811 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks May 15 14:56:11.476922 tar[1537]: linux-amd64/LICENSE May 15 14:56:11.476922 tar[1537]: linux-amd64/helm May 15 14:56:11.489856 jq[1543]: true May 15 14:56:11.512751 (ntainerd)[1559]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 15 14:56:11.531574 update_engine[1531]: I20250515 14:56:11.517237 1531 main.cc:92] Flatcar Update Engine starting May 15 14:56:11.530096 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 15 14:56:11.532278 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 15 14:56:11.541897 systemd[1]: Started update-engine.service - Update Engine. May 15 14:56:11.545755 update_engine[1531]: I20250515 14:56:11.543023 1531 update_check_scheduler.cc:74] Next update check in 4m1s May 15 14:56:11.556859 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 15 14:56:11.648563 kernel: EXT4-fs (vda9): resized filesystem to 15121403 May 15 14:56:11.688961 systemd[1]: extend-filesystems.service: Deactivated successfully. May 15 14:56:11.689621 extend-filesystems[1560]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 15 14:56:11.689621 extend-filesystems[1560]: old_desc_blocks = 1, new_desc_blocks = 8 May 15 14:56:11.689621 extend-filesystems[1560]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. May 15 14:56:11.689812 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 15 14:56:11.692699 extend-filesystems[1517]: Resized filesystem in /dev/vda9 May 15 14:56:11.692699 extend-filesystems[1517]: Found vdb May 15 14:56:11.721028 bash[1586]: Updated "/home/core/.ssh/authorized_keys" May 15 14:56:11.723136 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 15 14:56:11.739836 systemd[1]: Starting sshkeys.service... May 15 14:56:11.841176 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 15 14:56:11.844666 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 15 14:56:11.927182 coreos-metadata[1595]: May 15 14:56:11.926 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 15 14:56:11.988416 coreos-metadata[1595]: May 15 14:56:11.985 INFO Fetch successful May 15 14:56:11.997719 unknown[1595]: wrote ssh authorized keys file for user: core May 15 14:56:12.071682 update-ssh-keys[1600]: Updated "/home/core/.ssh/authorized_keys" May 15 14:56:12.066490 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 15 14:56:12.070457 systemd[1]: Finished sshkeys.service. May 15 14:56:12.100325 systemd-logind[1529]: New seat seat0. May 15 14:56:12.106542 systemd[1]: Started systemd-logind.service - User Login Management. May 15 14:56:12.172555 containerd[1559]: time="2025-05-15T14:56:12Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 15 14:56:12.176567 containerd[1559]: time="2025-05-15T14:56:12.175195847Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 15 14:56:12.189913 containerd[1559]: time="2025-05-15T14:56:12.189851225Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="17.759µs" May 15 14:56:12.190087 containerd[1559]: time="2025-05-15T14:56:12.190063239Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 15 14:56:12.190153 containerd[1559]: time="2025-05-15T14:56:12.190139437Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 15 14:56:12.190408 containerd[1559]: time="2025-05-15T14:56:12.190379601Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 15 14:56:12.190516 containerd[1559]: time="2025-05-15T14:56:12.190499161Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 15 14:56:12.190638 containerd[1559]: time="2025-05-15T14:56:12.190617005Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 15 14:56:12.190826 containerd[1559]: time="2025-05-15T14:56:12.190805120Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 15 14:56:12.191606 containerd[1559]: time="2025-05-15T14:56:12.191569910Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 15 14:56:12.192130 containerd[1559]: time="2025-05-15T14:56:12.192088339Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 15 14:56:12.192236 containerd[1559]: time="2025-05-15T14:56:12.192216187Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 15 14:56:12.192313 containerd[1559]: time="2025-05-15T14:56:12.192295672Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 15 14:56:12.192399 containerd[1559]: time="2025-05-15T14:56:12.192383977Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 15 14:56:12.192642 containerd[1559]: time="2025-05-15T14:56:12.192616217Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 15 14:56:12.193031 containerd[1559]: time="2025-05-15T14:56:12.193004155Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 15 14:56:12.193143 containerd[1559]: time="2025-05-15T14:56:12.193128142Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 15 14:56:12.193212 containerd[1559]: time="2025-05-15T14:56:12.193199451Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 15 14:56:12.193317 containerd[1559]: time="2025-05-15T14:56:12.193302418Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 15 14:56:12.193678 containerd[1559]: time="2025-05-15T14:56:12.193660581Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 15 14:56:12.193805 containerd[1559]: time="2025-05-15T14:56:12.193791109Z" level=info msg="metadata content store policy set" policy=shared May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.195974828Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196051662Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196085711Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196120378Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196135768Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196147300Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196160277Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196171600Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196182682Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196204196Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196215995Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196229763Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196357815Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 15 14:56:12.196568 containerd[1559]: time="2025-05-15T14:56:12.196377105Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196393774Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196404846Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196417775Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196427906Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196438842Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196449460Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196469318Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196480432Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 15 14:56:12.196932 containerd[1559]: time="2025-05-15T14:56:12.196493578Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 15 14:56:12.197295 containerd[1559]: time="2025-05-15T14:56:12.197258380Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 15 14:56:12.197399 containerd[1559]: time="2025-05-15T14:56:12.197381419Z" level=info msg="Start snapshots syncer" May 15 14:56:12.197493 containerd[1559]: time="2025-05-15T14:56:12.197475195Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 15 14:56:12.197984 containerd[1559]: time="2025-05-15T14:56:12.197939439Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 15 14:56:12.198561 containerd[1559]: time="2025-05-15T14:56:12.198213144Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 15 14:56:12.198561 containerd[1559]: time="2025-05-15T14:56:12.198352920Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 15 14:56:12.198561 containerd[1559]: time="2025-05-15T14:56:12.198494465Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 15 14:56:12.198561 containerd[1559]: time="2025-05-15T14:56:12.198518031Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 15 14:56:12.198723 containerd[1559]: time="2025-05-15T14:56:12.198529642Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 15 14:56:12.198823 containerd[1559]: time="2025-05-15T14:56:12.198806201Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 15 14:56:12.198885 containerd[1559]: time="2025-05-15T14:56:12.198874306Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 15 14:56:12.198929 containerd[1559]: time="2025-05-15T14:56:12.198920588Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 15 14:56:12.198976 containerd[1559]: time="2025-05-15T14:56:12.198966655Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 15 14:56:12.199077 containerd[1559]: time="2025-05-15T14:56:12.199065206Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 15 14:56:12.199126 containerd[1559]: time="2025-05-15T14:56:12.199117211Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 15 14:56:12.199183 containerd[1559]: time="2025-05-15T14:56:12.199172224Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 15 14:56:12.200798 containerd[1559]: time="2025-05-15T14:56:12.200766848Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 15 14:56:12.200927 containerd[1559]: time="2025-05-15T14:56:12.200904027Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 15 14:56:12.200994 containerd[1559]: time="2025-05-15T14:56:12.200981872Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 15 14:56:12.201043 containerd[1559]: time="2025-05-15T14:56:12.201031995Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 15 14:56:12.201088 containerd[1559]: time="2025-05-15T14:56:12.201075882Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 15 14:56:12.201183 containerd[1559]: time="2025-05-15T14:56:12.201164176Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 15 14:56:12.201241 containerd[1559]: time="2025-05-15T14:56:12.201230857Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 15 14:56:12.201297 containerd[1559]: time="2025-05-15T14:56:12.201286375Z" level=info msg="runtime interface created" May 15 14:56:12.201352 containerd[1559]: time="2025-05-15T14:56:12.201340367Z" level=info msg="created NRI interface" May 15 14:56:12.201418 containerd[1559]: time="2025-05-15T14:56:12.201405411Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 15 14:56:12.201469 containerd[1559]: time="2025-05-15T14:56:12.201460537Z" level=info msg="Connect containerd service" May 15 14:56:12.201632 containerd[1559]: time="2025-05-15T14:56:12.201608333Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 15 14:56:12.202824 containerd[1559]: time="2025-05-15T14:56:12.202790136Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 15 14:56:12.217119 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 14:56:12.240732 locksmithd[1572]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 15 14:56:12.252566 systemd-logind[1529]: Watching system buttons on /dev/input/event2 (Power Button) May 15 14:56:12.255725 systemd-logind[1529]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 15 14:56:12.402103 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 15 14:56:12.402637 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:12.405568 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 15 14:56:12.409861 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 15 14:56:12.634421 containerd[1559]: time="2025-05-15T14:56:12.634216987Z" level=info msg="Start subscribing containerd event" May 15 14:56:12.634421 containerd[1559]: time="2025-05-15T14:56:12.634323602Z" level=info msg="Start recovering state" May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634466525Z" level=info msg="Start event monitor" May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634490449Z" level=info msg="Start cni network conf syncer for default" May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634502094Z" level=info msg="Start streaming server" May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634514874Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634545496Z" level=info msg="runtime interface starting up..." May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634554879Z" level=info msg="starting plugins..." May 15 14:56:12.634653 containerd[1559]: time="2025-05-15T14:56:12.634574072Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 15 14:56:12.635326 containerd[1559]: time="2025-05-15T14:56:12.635291330Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 15 14:56:12.635529 containerd[1559]: time="2025-05-15T14:56:12.635485004Z" level=info msg=serving... address=/run/containerd/containerd.sock May 15 14:56:12.635730 systemd[1]: Started containerd.service - containerd container runtime. May 15 14:56:12.639636 containerd[1559]: time="2025-05-15T14:56:12.639593603Z" level=info msg="containerd successfully booted in 0.468129s" May 15 14:56:12.694563 kernel: EDAC MC: Ver: 3.0.0 May 15 14:56:12.714513 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 15 14:56:12.780150 sshd_keygen[1563]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 15 14:56:12.848626 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 15 14:56:12.857717 systemd[1]: Starting issuegen.service - Generate /run/issue... May 15 14:56:12.884436 systemd[1]: issuegen.service: Deactivated successfully. May 15 14:56:12.884908 systemd[1]: Finished issuegen.service - Generate /run/issue. May 15 14:56:12.887783 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 15 14:56:12.914537 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 15 14:56:12.915707 systemd-networkd[1431]: eth0: Gained IPv6LL May 15 14:56:12.920905 systemd[1]: Started getty@tty1.service - Getty on tty1. May 15 14:56:12.922829 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 15 14:56:12.923914 systemd[1]: Reached target getty.target - Login Prompts. May 15 14:56:12.926636 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 15 14:56:12.930903 systemd[1]: Reached target network-online.target - Network is Online. May 15 14:56:12.934484 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:12.938910 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 15 14:56:12.980338 systemd-networkd[1431]: eth1: Gained IPv6LL May 15 14:56:12.990615 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 15 14:56:13.095773 tar[1537]: linux-amd64/README.md May 15 14:56:13.128222 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 15 14:56:14.102280 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:14.110510 (kubelet)[1673]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 15 14:56:14.113906 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 15 14:56:14.114501 systemd[1]: Reached target multi-user.target - Multi-User System. May 15 14:56:14.116796 systemd[1]: Started sshd@0-24.199.109.202:22-139.178.68.195:50788.service - OpenSSH per-connection server daemon (139.178.68.195:50788). May 15 14:56:14.119399 systemd[1]: Startup finished in 3.699s (kernel) + 7.012s (initrd) + 6.889s (userspace) = 17.601s. May 15 14:56:14.233614 sshd[1675]: Accepted publickey for core from 139.178.68.195 port 50788 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:14.235741 sshd-session[1675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:14.249552 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 15 14:56:14.252543 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 15 14:56:14.270540 systemd-logind[1529]: New session 1 of user core. May 15 14:56:14.286699 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 15 14:56:14.291496 systemd[1]: Starting user@500.service - User Manager for UID 500... May 15 14:56:14.310456 (systemd)[1688]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 15 14:56:14.315832 systemd-logind[1529]: New session c1 of user core. May 15 14:56:14.518844 systemd[1688]: Queued start job for default target default.target. May 15 14:56:14.524628 systemd[1688]: Created slice app.slice - User Application Slice. May 15 14:56:14.524682 systemd[1688]: Reached target paths.target - Paths. May 15 14:56:14.524746 systemd[1688]: Reached target timers.target - Timers. May 15 14:56:14.527712 systemd[1688]: Starting dbus.socket - D-Bus User Message Bus Socket... May 15 14:56:14.552384 systemd[1688]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 15 14:56:14.553332 systemd[1688]: Reached target sockets.target - Sockets. May 15 14:56:14.553396 systemd[1688]: Reached target basic.target - Basic System. May 15 14:56:14.553437 systemd[1688]: Reached target default.target - Main User Target. May 15 14:56:14.553478 systemd[1688]: Startup finished in 223ms. May 15 14:56:14.554113 systemd[1]: Started user@500.service - User Manager for UID 500. May 15 14:56:14.562254 systemd[1]: Started session-1.scope - Session 1 of User core. May 15 14:56:14.643816 systemd[1]: Started sshd@1-24.199.109.202:22-139.178.68.195:50790.service - OpenSSH per-connection server daemon (139.178.68.195:50790). May 15 14:56:14.730567 sshd[1699]: Accepted publickey for core from 139.178.68.195 port 50790 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:14.732965 sshd-session[1699]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:14.740945 systemd-logind[1529]: New session 2 of user core. May 15 14:56:14.745859 systemd[1]: Started session-2.scope - Session 2 of User core. May 15 14:56:14.810652 sshd[1702]: Connection closed by 139.178.68.195 port 50790 May 15 14:56:14.814266 sshd-session[1699]: pam_unix(sshd:session): session closed for user core May 15 14:56:14.823835 systemd[1]: sshd@1-24.199.109.202:22-139.178.68.195:50790.service: Deactivated successfully. May 15 14:56:14.827410 systemd[1]: session-2.scope: Deactivated successfully. May 15 14:56:14.829245 systemd-logind[1529]: Session 2 logged out. Waiting for processes to exit. May 15 14:56:14.838928 systemd[1]: Started sshd@2-24.199.109.202:22-139.178.68.195:50800.service - OpenSSH per-connection server daemon (139.178.68.195:50800). May 15 14:56:14.841355 systemd-logind[1529]: Removed session 2. May 15 14:56:14.912382 sshd[1708]: Accepted publickey for core from 139.178.68.195 port 50800 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:14.914815 sshd-session[1708]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:14.925889 systemd-logind[1529]: New session 3 of user core. May 15 14:56:14.928132 kubelet[1673]: E0515 14:56:14.928039 1673 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 15 14:56:14.928756 systemd[1]: Started session-3.scope - Session 3 of User core. May 15 14:56:14.935389 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 15 14:56:14.935909 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 15 14:56:14.936626 systemd[1]: kubelet.service: Consumed 1.361s CPU time, 251.6M memory peak. May 15 14:56:14.990581 sshd[1711]: Connection closed by 139.178.68.195 port 50800 May 15 14:56:14.991214 sshd-session[1708]: pam_unix(sshd:session): session closed for user core May 15 14:56:15.005282 systemd[1]: sshd@2-24.199.109.202:22-139.178.68.195:50800.service: Deactivated successfully. May 15 14:56:15.008204 systemd[1]: session-3.scope: Deactivated successfully. May 15 14:56:15.010717 systemd-logind[1529]: Session 3 logged out. Waiting for processes to exit. May 15 14:56:15.014292 systemd[1]: Started sshd@3-24.199.109.202:22-139.178.68.195:50816.service - OpenSSH per-connection server daemon (139.178.68.195:50816). May 15 14:56:15.015724 systemd-logind[1529]: Removed session 3. May 15 14:56:15.086060 sshd[1717]: Accepted publickey for core from 139.178.68.195 port 50816 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:15.087312 sshd-session[1717]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:15.095379 systemd-logind[1529]: New session 4 of user core. May 15 14:56:15.108889 systemd[1]: Started session-4.scope - Session 4 of User core. May 15 14:56:15.172959 sshd[1719]: Connection closed by 139.178.68.195 port 50816 May 15 14:56:15.173661 sshd-session[1717]: pam_unix(sshd:session): session closed for user core May 15 14:56:15.188047 systemd[1]: sshd@3-24.199.109.202:22-139.178.68.195:50816.service: Deactivated successfully. May 15 14:56:15.190594 systemd[1]: session-4.scope: Deactivated successfully. May 15 14:56:15.191889 systemd-logind[1529]: Session 4 logged out. Waiting for processes to exit. May 15 14:56:15.195714 systemd[1]: Started sshd@4-24.199.109.202:22-139.178.68.195:50828.service - OpenSSH per-connection server daemon (139.178.68.195:50828). May 15 14:56:15.197197 systemd-logind[1529]: Removed session 4. May 15 14:56:15.264721 sshd[1725]: Accepted publickey for core from 139.178.68.195 port 50828 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:15.266889 sshd-session[1725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:15.273892 systemd-logind[1529]: New session 5 of user core. May 15 14:56:15.283825 systemd[1]: Started session-5.scope - Session 5 of User core. May 15 14:56:15.356221 sudo[1728]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 15 14:56:15.357122 sudo[1728]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 14:56:15.372437 sudo[1728]: pam_unix(sudo:session): session closed for user root May 15 14:56:15.376602 sshd[1727]: Connection closed by 139.178.68.195 port 50828 May 15 14:56:15.377573 sshd-session[1725]: pam_unix(sshd:session): session closed for user core May 15 14:56:15.393358 systemd[1]: sshd@4-24.199.109.202:22-139.178.68.195:50828.service: Deactivated successfully. May 15 14:56:15.395694 systemd[1]: session-5.scope: Deactivated successfully. May 15 14:56:15.396897 systemd-logind[1529]: Session 5 logged out. Waiting for processes to exit. May 15 14:56:15.402292 systemd[1]: Started sshd@5-24.199.109.202:22-139.178.68.195:50834.service - OpenSSH per-connection server daemon (139.178.68.195:50834). May 15 14:56:15.403664 systemd-logind[1529]: Removed session 5. May 15 14:56:15.478633 sshd[1734]: Accepted publickey for core from 139.178.68.195 port 50834 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:15.480409 sshd-session[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:15.486489 systemd-logind[1529]: New session 6 of user core. May 15 14:56:15.493907 systemd[1]: Started session-6.scope - Session 6 of User core. May 15 14:56:15.555911 sudo[1738]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 15 14:56:15.556364 sudo[1738]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 14:56:15.563130 sudo[1738]: pam_unix(sudo:session): session closed for user root May 15 14:56:15.570587 sudo[1737]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 15 14:56:15.571263 sudo[1737]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 14:56:15.589135 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 15 14:56:15.646268 augenrules[1760]: No rules May 15 14:56:15.646290 systemd[1]: audit-rules.service: Deactivated successfully. May 15 14:56:15.646593 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 15 14:56:15.648944 sudo[1737]: pam_unix(sudo:session): session closed for user root May 15 14:56:15.652610 sshd[1736]: Connection closed by 139.178.68.195 port 50834 May 15 14:56:15.653458 sshd-session[1734]: pam_unix(sshd:session): session closed for user core May 15 14:56:15.664868 systemd[1]: sshd@5-24.199.109.202:22-139.178.68.195:50834.service: Deactivated successfully. May 15 14:56:15.667571 systemd[1]: session-6.scope: Deactivated successfully. May 15 14:56:15.668753 systemd-logind[1529]: Session 6 logged out. Waiting for processes to exit. May 15 14:56:15.673410 systemd[1]: Started sshd@6-24.199.109.202:22-139.178.68.195:50840.service - OpenSSH per-connection server daemon (139.178.68.195:50840). May 15 14:56:15.674642 systemd-logind[1529]: Removed session 6. May 15 14:56:15.745267 sshd[1769]: Accepted publickey for core from 139.178.68.195 port 50840 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:56:15.746991 sshd-session[1769]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:56:15.754027 systemd-logind[1529]: New session 7 of user core. May 15 14:56:15.759847 systemd[1]: Started session-7.scope - Session 7 of User core. May 15 14:56:15.822630 sudo[1772]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 15 14:56:15.823066 sudo[1772]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 15 14:56:16.476607 systemd[1]: Starting docker.service - Docker Application Container Engine... May 15 14:56:16.494295 (dockerd)[1790]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 15 14:56:16.946551 dockerd[1790]: time="2025-05-15T14:56:16.946352957Z" level=info msg="Starting up" May 15 14:56:16.949249 dockerd[1790]: time="2025-05-15T14:56:16.949189781Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 15 14:56:16.988802 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1553138368-merged.mount: Deactivated successfully. May 15 14:56:17.060567 dockerd[1790]: time="2025-05-15T14:56:17.060427511Z" level=info msg="Loading containers: start." May 15 14:56:17.073754 kernel: Initializing XFRM netlink socket May 15 14:56:17.380933 systemd-networkd[1431]: docker0: Link UP May 15 14:56:17.384269 dockerd[1790]: time="2025-05-15T14:56:17.384158986Z" level=info msg="Loading containers: done." May 15 14:56:17.403702 dockerd[1790]: time="2025-05-15T14:56:17.403173659Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 15 14:56:17.403702 dockerd[1790]: time="2025-05-15T14:56:17.403304814Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 15 14:56:17.403702 dockerd[1790]: time="2025-05-15T14:56:17.403475442Z" level=info msg="Initializing buildkit" May 15 14:56:17.431724 dockerd[1790]: time="2025-05-15T14:56:17.431670294Z" level=info msg="Completed buildkit initialization" May 15 14:56:17.436677 dockerd[1790]: time="2025-05-15T14:56:17.436625099Z" level=info msg="Daemon has completed initialization" May 15 14:56:17.436978 systemd[1]: Started docker.service - Docker Application Container Engine. May 15 14:56:17.437194 dockerd[1790]: time="2025-05-15T14:56:17.437057175Z" level=info msg="API listen on /run/docker.sock" May 15 14:56:18.327750 containerd[1559]: time="2025-05-15T14:56:18.327640344Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\"" May 15 14:56:18.972208 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4041874617.mount: Deactivated successfully. May 15 14:56:20.571657 containerd[1559]: time="2025-05-15T14:56:20.571592473Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:20.572900 containerd[1559]: time="2025-05-15T14:56:20.572495519Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.4: active requests=0, bytes read=28682879" May 15 14:56:20.573596 containerd[1559]: time="2025-05-15T14:56:20.573559678Z" level=info msg="ImageCreate event name:\"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:20.576194 containerd[1559]: time="2025-05-15T14:56:20.576146430Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:20.577815 containerd[1559]: time="2025-05-15T14:56:20.577388427Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.4\" with image id \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\", size \"28679679\" in 2.249648281s" May 15 14:56:20.577815 containerd[1559]: time="2025-05-15T14:56:20.577441961Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\" returns image reference \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\"" May 15 14:56:20.578729 containerd[1559]: time="2025-05-15T14:56:20.578692965Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\"" May 15 14:56:22.470558 containerd[1559]: time="2025-05-15T14:56:22.469890014Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:22.472506 containerd[1559]: time="2025-05-15T14:56:22.472459291Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.4: active requests=0, bytes read=24779589" May 15 14:56:22.473653 containerd[1559]: time="2025-05-15T14:56:22.472995095Z" level=info msg="ImageCreate event name:\"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:22.478556 containerd[1559]: time="2025-05-15T14:56:22.478094854Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:22.479482 containerd[1559]: time="2025-05-15T14:56:22.479351369Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.4\" with image id \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\", size \"26267962\" in 1.90062334s" May 15 14:56:22.479482 containerd[1559]: time="2025-05-15T14:56:22.479402872Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\" returns image reference \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\"" May 15 14:56:22.486555 containerd[1559]: time="2025-05-15T14:56:22.486359757Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\"" May 15 14:56:23.851099 containerd[1559]: time="2025-05-15T14:56:23.851024418Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:23.852616 containerd[1559]: time="2025-05-15T14:56:23.852555275Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.4: active requests=0, bytes read=19169938" May 15 14:56:23.855559 containerd[1559]: time="2025-05-15T14:56:23.854799054Z" level=info msg="ImageCreate event name:\"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:23.859153 containerd[1559]: time="2025-05-15T14:56:23.858998904Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:23.860635 containerd[1559]: time="2025-05-15T14:56:23.860588342Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.4\" with image id \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\", size \"20658329\" in 1.374159382s" May 15 14:56:23.860802 containerd[1559]: time="2025-05-15T14:56:23.860785509Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\" returns image reference \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\"" May 15 14:56:23.862115 containerd[1559]: time="2025-05-15T14:56:23.862077835Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\"" May 15 14:56:24.953509 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1964237425.mount: Deactivated successfully. May 15 14:56:24.955891 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 15 14:56:24.958823 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:25.159990 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:25.176615 (kubelet)[2081]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 15 14:56:25.270892 kubelet[2081]: E0515 14:56:25.270782 2081 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 15 14:56:25.275927 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 15 14:56:25.276130 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 15 14:56:25.277723 systemd[1]: kubelet.service: Consumed 246ms CPU time, 103.7M memory peak. May 15 14:56:25.709132 containerd[1559]: time="2025-05-15T14:56:25.708628148Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:25.709932 containerd[1559]: time="2025-05-15T14:56:25.709767564Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.4: active requests=0, bytes read=30917856" May 15 14:56:25.710580 containerd[1559]: time="2025-05-15T14:56:25.710540431Z" level=info msg="ImageCreate event name:\"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:25.712462 containerd[1559]: time="2025-05-15T14:56:25.712422381Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:25.713556 containerd[1559]: time="2025-05-15T14:56:25.713499939Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.4\" with image id \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\", repo tag \"registry.k8s.io/kube-proxy:v1.32.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\", size \"30916875\" in 1.851362441s" May 15 14:56:25.714125 containerd[1559]: time="2025-05-15T14:56:25.713673410Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\" returns image reference \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\"" May 15 14:56:25.714935 containerd[1559]: time="2025-05-15T14:56:25.714790280Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 15 14:56:25.716788 systemd-resolved[1401]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. May 15 14:56:26.278447 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1219706153.mount: Deactivated successfully. May 15 14:56:27.311552 containerd[1559]: time="2025-05-15T14:56:27.310549585Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:27.313009 containerd[1559]: time="2025-05-15T14:56:27.312961726Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" May 15 14:56:27.313778 containerd[1559]: time="2025-05-15T14:56:27.313746093Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:27.318269 containerd[1559]: time="2025-05-15T14:56:27.318215596Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.603389491s" May 15 14:56:27.318533 containerd[1559]: time="2025-05-15T14:56:27.318510272Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 15 14:56:27.318721 containerd[1559]: time="2025-05-15T14:56:27.318470554Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:27.320236 containerd[1559]: time="2025-05-15T14:56:27.320046408Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 15 14:56:27.828347 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4009091464.mount: Deactivated successfully. May 15 14:56:27.834565 containerd[1559]: time="2025-05-15T14:56:27.833657286Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 14:56:27.835288 containerd[1559]: time="2025-05-15T14:56:27.835215310Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 15 14:56:27.836141 containerd[1559]: time="2025-05-15T14:56:27.836099532Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 14:56:27.838877 containerd[1559]: time="2025-05-15T14:56:27.838827050Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 15 14:56:27.839936 containerd[1559]: time="2025-05-15T14:56:27.839886168Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 519.444568ms" May 15 14:56:27.839936 containerd[1559]: time="2025-05-15T14:56:27.839936660Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 15 14:56:27.841370 containerd[1559]: time="2025-05-15T14:56:27.841315326Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" May 15 14:56:28.315347 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1590867116.mount: Deactivated successfully. May 15 14:56:28.787772 systemd-resolved[1401]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. May 15 14:56:30.346900 containerd[1559]: time="2025-05-15T14:56:30.345372784Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:30.346900 containerd[1559]: time="2025-05-15T14:56:30.346544731Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551360" May 15 14:56:30.346900 containerd[1559]: time="2025-05-15T14:56:30.346822683Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:30.350556 containerd[1559]: time="2025-05-15T14:56:30.350477181Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:30.352389 containerd[1559]: time="2025-05-15T14:56:30.352324866Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.510960433s" May 15 14:56:30.352389 containerd[1559]: time="2025-05-15T14:56:30.352387817Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" May 15 14:56:33.973381 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:33.973761 systemd[1]: kubelet.service: Consumed 246ms CPU time, 103.7M memory peak. May 15 14:56:33.976410 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:34.020898 systemd[1]: Reload requested from client PID 2225 ('systemctl') (unit session-7.scope)... May 15 14:56:34.020952 systemd[1]: Reloading... May 15 14:56:34.188564 zram_generator::config[2264]: No configuration found. May 15 14:56:34.316656 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 14:56:34.460007 systemd[1]: Reloading finished in 438 ms. May 15 14:56:34.517186 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 15 14:56:34.517334 systemd[1]: kubelet.service: Failed with result 'signal'. May 15 14:56:34.517802 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:34.517882 systemd[1]: kubelet.service: Consumed 136ms CPU time, 91.8M memory peak. May 15 14:56:34.520507 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:34.690126 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:34.701317 (kubelet)[2322]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 15 14:56:34.769556 kubelet[2322]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 14:56:34.770043 kubelet[2322]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 15 14:56:34.770104 kubelet[2322]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 14:56:34.770362 kubelet[2322]: I0515 14:56:34.770315 2322 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 15 14:56:35.148207 kubelet[2322]: I0515 14:56:35.147730 2322 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 15 14:56:35.148207 kubelet[2322]: I0515 14:56:35.147777 2322 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 15 14:56:35.148798 kubelet[2322]: I0515 14:56:35.148769 2322 server.go:954] "Client rotation is on, will bootstrap in background" May 15 14:56:35.190213 kubelet[2322]: I0515 14:56:35.190166 2322 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 15 14:56:35.192620 kubelet[2322]: E0515 14:56:35.192312 2322 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://24.199.109.202:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:35.206277 kubelet[2322]: I0515 14:56:35.206170 2322 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 15 14:56:35.211865 kubelet[2322]: I0515 14:56:35.211823 2322 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 15 14:56:35.213556 kubelet[2322]: I0515 14:56:35.213453 2322 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 15 14:56:35.213831 kubelet[2322]: I0515 14:56:35.213548 2322 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4334.0.0-a-c1acadba3f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 15 14:56:35.213964 kubelet[2322]: I0515 14:56:35.213836 2322 topology_manager.go:138] "Creating topology manager with none policy" May 15 14:56:35.213964 kubelet[2322]: I0515 14:56:35.213851 2322 container_manager_linux.go:304] "Creating device plugin manager" May 15 14:56:35.216359 kubelet[2322]: I0515 14:56:35.216283 2322 state_mem.go:36] "Initialized new in-memory state store" May 15 14:56:35.220387 kubelet[2322]: I0515 14:56:35.220254 2322 kubelet.go:446] "Attempting to sync node with API server" May 15 14:56:35.220387 kubelet[2322]: I0515 14:56:35.220296 2322 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 15 14:56:35.220387 kubelet[2322]: I0515 14:56:35.220334 2322 kubelet.go:352] "Adding apiserver pod source" May 15 14:56:35.220387 kubelet[2322]: I0515 14:56:35.220349 2322 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 15 14:56:35.227884 kubelet[2322]: W0515 14:56:35.227591 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://24.199.109.202:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-c1acadba3f&limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:35.228207 kubelet[2322]: E0515 14:56:35.228157 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://24.199.109.202:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-c1acadba3f&limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:35.230371 kubelet[2322]: I0515 14:56:35.230312 2322 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 15 14:56:35.236561 kubelet[2322]: I0515 14:56:35.234118 2322 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 15 14:56:35.239542 kubelet[2322]: W0515 14:56:35.238968 2322 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 15 14:56:35.239990 kubelet[2322]: I0515 14:56:35.239960 2322 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 15 14:56:35.240106 kubelet[2322]: I0515 14:56:35.240056 2322 server.go:1287] "Started kubelet" May 15 14:56:35.242977 kubelet[2322]: W0515 14:56:35.242910 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://24.199.109.202:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:35.243222 kubelet[2322]: E0515 14:56:35.243192 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://24.199.109.202:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:35.243545 kubelet[2322]: I0515 14:56:35.243442 2322 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 15 14:56:35.244095 kubelet[2322]: I0515 14:56:35.244060 2322 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 15 14:56:35.244210 kubelet[2322]: I0515 14:56:35.243984 2322 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 15 14:56:35.248469 kubelet[2322]: I0515 14:56:35.248425 2322 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 15 14:56:35.252717 kubelet[2322]: I0515 14:56:35.252685 2322 server.go:490] "Adding debug handlers to kubelet server" May 15 14:56:35.253724 kubelet[2322]: I0515 14:56:35.253692 2322 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 15 14:56:35.257570 kubelet[2322]: E0515 14:56:35.257466 2322 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" May 15 14:56:35.258037 kubelet[2322]: I0515 14:56:35.257961 2322 volume_manager.go:297] "Starting Kubelet Volume Manager" May 15 14:56:35.258590 kubelet[2322]: E0515 14:56:35.258484 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.199.109.202:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-c1acadba3f?timeout=10s\": dial tcp 24.199.109.202:6443: connect: connection refused" interval="200ms" May 15 14:56:35.260548 kubelet[2322]: E0515 14:56:35.258756 2322 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://24.199.109.202:6443/api/v1/namespaces/default/events\": dial tcp 24.199.109.202:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4334.0.0-a-c1acadba3f.183fbb2ffddc718e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4334.0.0-a-c1acadba3f,UID:ci-4334.0.0-a-c1acadba3f,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4334.0.0-a-c1acadba3f,},FirstTimestamp:2025-05-15 14:56:35.239981454 +0000 UTC m=+0.533873596,LastTimestamp:2025-05-15 14:56:35.239981454 +0000 UTC m=+0.533873596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4334.0.0-a-c1acadba3f,}" May 15 14:56:35.261135 kubelet[2322]: W0515 14:56:35.261080 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://24.199.109.202:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:35.261281 kubelet[2322]: E0515 14:56:35.261250 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://24.199.109.202:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:35.261379 kubelet[2322]: I0515 14:56:35.261369 2322 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 15 14:56:35.261531 kubelet[2322]: I0515 14:56:35.261489 2322 reconciler.go:26] "Reconciler: start to sync state" May 15 14:56:35.280891 kubelet[2322]: E0515 14:56:35.280864 2322 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 15 14:56:35.290156 kubelet[2322]: I0515 14:56:35.290119 2322 factory.go:221] Registration of the containerd container factory successfully May 15 14:56:35.293573 kubelet[2322]: I0515 14:56:35.292580 2322 factory.go:221] Registration of the systemd container factory successfully May 15 14:56:35.293573 kubelet[2322]: I0515 14:56:35.292781 2322 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 15 14:56:35.305336 kubelet[2322]: I0515 14:56:35.290330 2322 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 15 14:56:35.310216 kubelet[2322]: I0515 14:56:35.309507 2322 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 15 14:56:35.310216 kubelet[2322]: I0515 14:56:35.309572 2322 status_manager.go:227] "Starting to sync pod status with apiserver" May 15 14:56:35.310216 kubelet[2322]: I0515 14:56:35.309600 2322 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 15 14:56:35.310216 kubelet[2322]: I0515 14:56:35.309610 2322 kubelet.go:2388] "Starting kubelet main sync loop" May 15 14:56:35.310216 kubelet[2322]: E0515 14:56:35.309680 2322 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 15 14:56:35.314510 kubelet[2322]: W0515 14:56:35.314437 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://24.199.109.202:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:35.315963 kubelet[2322]: E0515 14:56:35.315920 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://24.199.109.202:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:35.318049 kubelet[2322]: I0515 14:56:35.318022 2322 cpu_manager.go:221] "Starting CPU manager" policy="none" May 15 14:56:35.318049 kubelet[2322]: I0515 14:56:35.318038 2322 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 15 14:56:35.318049 kubelet[2322]: I0515 14:56:35.318060 2322 state_mem.go:36] "Initialized new in-memory state store" May 15 14:56:35.321358 kubelet[2322]: I0515 14:56:35.321323 2322 policy_none.go:49] "None policy: Start" May 15 14:56:35.321358 kubelet[2322]: I0515 14:56:35.321363 2322 memory_manager.go:186] "Starting memorymanager" policy="None" May 15 14:56:35.321517 kubelet[2322]: I0515 14:56:35.321385 2322 state_mem.go:35] "Initializing new in-memory state store" May 15 14:56:35.331681 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 15 14:56:35.352721 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 15 14:56:35.357560 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 15 14:56:35.358104 kubelet[2322]: E0515 14:56:35.357914 2322 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" May 15 14:56:35.363509 kubelet[2322]: I0515 14:56:35.363479 2322 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 15 14:56:35.364392 kubelet[2322]: I0515 14:56:35.364281 2322 eviction_manager.go:189] "Eviction manager: starting control loop" May 15 14:56:35.365714 kubelet[2322]: I0515 14:56:35.365659 2322 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 15 14:56:35.366040 kubelet[2322]: I0515 14:56:35.365984 2322 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 15 14:56:35.369218 kubelet[2322]: E0515 14:56:35.369194 2322 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 15 14:56:35.369335 kubelet[2322]: E0515 14:56:35.369236 2322 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4334.0.0-a-c1acadba3f\" not found" May 15 14:56:35.422371 systemd[1]: Created slice kubepods-burstable-pod4b35d03f5ff6f301272d281a76f73df9.slice - libcontainer container kubepods-burstable-pod4b35d03f5ff6f301272d281a76f73df9.slice. May 15 14:56:35.444323 kubelet[2322]: E0515 14:56:35.443881 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.447048 systemd[1]: Created slice kubepods-burstable-pod3bb4783d2c516c9c05f69ae7d3e1d3cd.slice - libcontainer container kubepods-burstable-pod3bb4783d2c516c9c05f69ae7d3e1d3cd.slice. May 15 14:56:35.450108 kubelet[2322]: E0515 14:56:35.450081 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.454550 systemd[1]: Created slice kubepods-burstable-poddc64a222dc9b3fd1d547417b64b8c506.slice - libcontainer container kubepods-burstable-poddc64a222dc9b3fd1d547417b64b8c506.slice. May 15 14:56:35.457781 kubelet[2322]: E0515 14:56:35.457624 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.459100 kubelet[2322]: E0515 14:56:35.459047 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.199.109.202:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-c1acadba3f?timeout=10s\": dial tcp 24.199.109.202:6443: connect: connection refused" interval="400ms" May 15 14:56:35.470970 kubelet[2322]: I0515 14:56:35.470920 2322 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.473964 kubelet[2322]: E0515 14:56:35.473903 2322 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.199.109.202:6443/api/v1/nodes\": dial tcp 24.199.109.202:6443: connect: connection refused" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.562976 kubelet[2322]: I0515 14:56:35.562707 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.562976 kubelet[2322]: I0515 14:56:35.562760 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-k8s-certs\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.562976 kubelet[2322]: I0515 14:56:35.562785 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.562976 kubelet[2322]: I0515 14:56:35.562800 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-flexvolume-dir\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.562976 kubelet[2322]: I0515 14:56:35.562816 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-k8s-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.563248 kubelet[2322]: I0515 14:56:35.562830 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-kubeconfig\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.563248 kubelet[2322]: I0515 14:56:35.562851 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dc64a222dc9b3fd1d547417b64b8c506-kubeconfig\") pod \"kube-scheduler-ci-4334.0.0-a-c1acadba3f\" (UID: \"dc64a222dc9b3fd1d547417b64b8c506\") " pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.563248 kubelet[2322]: I0515 14:56:35.562871 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-ca-certs\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.563248 kubelet[2322]: I0515 14:56:35.562891 2322 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-ca-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.676704 kubelet[2322]: I0515 14:56:35.676049 2322 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.676704 kubelet[2322]: E0515 14:56:35.676516 2322 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.199.109.202:6443/api/v1/nodes\": dial tcp 24.199.109.202:6443: connect: connection refused" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:35.745655 kubelet[2322]: E0515 14:56:35.745575 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:35.746735 containerd[1559]: time="2025-05-15T14:56:35.746569685Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4334.0.0-a-c1acadba3f,Uid:4b35d03f5ff6f301272d281a76f73df9,Namespace:kube-system,Attempt:0,}" May 15 14:56:35.751965 kubelet[2322]: E0515 14:56:35.751910 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:35.760289 kubelet[2322]: E0515 14:56:35.759790 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:35.763064 containerd[1559]: time="2025-05-15T14:56:35.762998291Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4334.0.0-a-c1acadba3f,Uid:dc64a222dc9b3fd1d547417b64b8c506,Namespace:kube-system,Attempt:0,}" May 15 14:56:35.763715 containerd[1559]: time="2025-05-15T14:56:35.763598745Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4334.0.0-a-c1acadba3f,Uid:3bb4783d2c516c9c05f69ae7d3e1d3cd,Namespace:kube-system,Attempt:0,}" May 15 14:56:35.859809 kubelet[2322]: E0515 14:56:35.859764 2322 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://24.199.109.202:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4334.0.0-a-c1acadba3f?timeout=10s\": dial tcp 24.199.109.202:6443: connect: connection refused" interval="800ms" May 15 14:56:35.922675 containerd[1559]: time="2025-05-15T14:56:35.922458729Z" level=info msg="connecting to shim d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04" address="unix:///run/containerd/s/4f12f18b6a1af1a35fa2c0f6bf8c7c15fc253e1ab5ff0ec6ec24eaa394887724" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:35.926604 containerd[1559]: time="2025-05-15T14:56:35.925779256Z" level=info msg="connecting to shim a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c" address="unix:///run/containerd/s/d44808c648ce68c2bc546beefca3e82cab744d7127bfbf2bd13f23829dd275e7" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:35.927131 containerd[1559]: time="2025-05-15T14:56:35.927002881Z" level=info msg="connecting to shim 652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231" address="unix:///run/containerd/s/de34215cb5906cb2b46dd567ae9c7f34c6c0c73fa131ea62c6a84ac05c8711db" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:36.061803 systemd[1]: Started cri-containerd-652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231.scope - libcontainer container 652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231. May 15 14:56:36.063537 systemd[1]: Started cri-containerd-a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c.scope - libcontainer container a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c. May 15 14:56:36.066200 systemd[1]: Started cri-containerd-d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04.scope - libcontainer container d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04. May 15 14:56:36.078282 kubelet[2322]: I0515 14:56:36.078251 2322 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:36.079118 kubelet[2322]: E0515 14:56:36.078696 2322 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://24.199.109.202:6443/api/v1/nodes\": dial tcp 24.199.109.202:6443: connect: connection refused" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:36.079118 kubelet[2322]: W0515 14:56:36.078931 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://24.199.109.202:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:36.079118 kubelet[2322]: E0515 14:56:36.079013 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://24.199.109.202:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:36.086565 kubelet[2322]: W0515 14:56:36.083753 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://24.199.109.202:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-c1acadba3f&limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:36.086565 kubelet[2322]: E0515 14:56:36.083826 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://24.199.109.202:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4334.0.0-a-c1acadba3f&limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:36.205731 containerd[1559]: time="2025-05-15T14:56:36.205495601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4334.0.0-a-c1acadba3f,Uid:4b35d03f5ff6f301272d281a76f73df9,Namespace:kube-system,Attempt:0,} returns sandbox id \"a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c\"" May 15 14:56:36.209561 kubelet[2322]: E0515 14:56:36.207550 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:36.216047 containerd[1559]: time="2025-05-15T14:56:36.215957346Z" level=info msg="CreateContainer within sandbox \"a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 15 14:56:36.221594 containerd[1559]: time="2025-05-15T14:56:36.221013416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4334.0.0-a-c1acadba3f,Uid:3bb4783d2c516c9c05f69ae7d3e1d3cd,Namespace:kube-system,Attempt:0,} returns sandbox id \"652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231\"" May 15 14:56:36.227859 kubelet[2322]: E0515 14:56:36.227819 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:36.235797 containerd[1559]: time="2025-05-15T14:56:36.235737108Z" level=info msg="CreateContainer within sandbox \"652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 15 14:56:36.246005 containerd[1559]: time="2025-05-15T14:56:36.245576873Z" level=info msg="Container d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:36.246005 containerd[1559]: time="2025-05-15T14:56:36.245731825Z" level=info msg="Container b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:36.250161 containerd[1559]: time="2025-05-15T14:56:36.250124523Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4334.0.0-a-c1acadba3f,Uid:dc64a222dc9b3fd1d547417b64b8c506,Namespace:kube-system,Attempt:0,} returns sandbox id \"d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04\"" May 15 14:56:36.251563 kubelet[2322]: E0515 14:56:36.251536 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:36.255036 containerd[1559]: time="2025-05-15T14:56:36.255001743Z" level=info msg="CreateContainer within sandbox \"d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 15 14:56:36.257169 containerd[1559]: time="2025-05-15T14:56:36.257124466Z" level=info msg="CreateContainer within sandbox \"652131362d2f9a764673ab751b8cd952386c5c2b83a8b9ecbf58c75435938231\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c\"" May 15 14:56:36.259277 containerd[1559]: time="2025-05-15T14:56:36.259236644Z" level=info msg="CreateContainer within sandbox \"a1403c1fea8f7de3ab6538f1bb8775ad118b26a50b9471946e0002369a98956c\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad\"" May 15 14:56:36.260387 containerd[1559]: time="2025-05-15T14:56:36.260354320Z" level=info msg="StartContainer for \"b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad\"" May 15 14:56:36.260711 containerd[1559]: time="2025-05-15T14:56:36.260638103Z" level=info msg="StartContainer for \"d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c\"" May 15 14:56:36.261932 containerd[1559]: time="2025-05-15T14:56:36.261900409Z" level=info msg="connecting to shim b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad" address="unix:///run/containerd/s/d44808c648ce68c2bc546beefca3e82cab744d7127bfbf2bd13f23829dd275e7" protocol=ttrpc version=3 May 15 14:56:36.262549 containerd[1559]: time="2025-05-15T14:56:36.262477510Z" level=info msg="connecting to shim d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c" address="unix:///run/containerd/s/de34215cb5906cb2b46dd567ae9c7f34c6c0c73fa131ea62c6a84ac05c8711db" protocol=ttrpc version=3 May 15 14:56:36.269042 containerd[1559]: time="2025-05-15T14:56:36.268955140Z" level=info msg="Container f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:36.275357 containerd[1559]: time="2025-05-15T14:56:36.275306263Z" level=info msg="CreateContainer within sandbox \"d8c1d49070a235b08e326e41fa8cff3033023d09477c1a5598e5b6a6144f7d04\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea\"" May 15 14:56:36.276671 containerd[1559]: time="2025-05-15T14:56:36.275946162Z" level=info msg="StartContainer for \"f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea\"" May 15 14:56:36.277611 containerd[1559]: time="2025-05-15T14:56:36.277576219Z" level=info msg="connecting to shim f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea" address="unix:///run/containerd/s/4f12f18b6a1af1a35fa2c0f6bf8c7c15fc253e1ab5ff0ec6ec24eaa394887724" protocol=ttrpc version=3 May 15 14:56:36.313820 systemd[1]: Started cri-containerd-b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad.scope - libcontainer container b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad. May 15 14:56:36.316219 systemd[1]: Started cri-containerd-d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c.scope - libcontainer container d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c. May 15 14:56:36.325127 systemd[1]: Started cri-containerd-f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea.scope - libcontainer container f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea. May 15 14:56:36.436143 containerd[1559]: time="2025-05-15T14:56:36.436099851Z" level=info msg="StartContainer for \"d32bdcb6e5c057e2404953d8af9e3ffb63105c4e3a12d42da12d78473c919c3c\" returns successfully" May 15 14:56:36.460377 containerd[1559]: time="2025-05-15T14:56:36.459568547Z" level=info msg="StartContainer for \"b7388b520816e0286b1c9a5d45a3e1580c20cfcc1765bc587b8fc5f519dbb1ad\" returns successfully" May 15 14:56:36.480122 containerd[1559]: time="2025-05-15T14:56:36.480066218Z" level=info msg="StartContainer for \"f2b75c9773a7ed42aa2791a205eea34d296895c0e78e06d3dd83400adbef76ea\" returns successfully" May 15 14:56:36.494817 kubelet[2322]: W0515 14:56:36.494699 2322 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://24.199.109.202:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 24.199.109.202:6443: connect: connection refused May 15 14:56:36.495081 kubelet[2322]: E0515 14:56:36.495014 2322 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://24.199.109.202:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 24.199.109.202:6443: connect: connection refused" logger="UnhandledError" May 15 14:56:36.880924 kubelet[2322]: I0515 14:56:36.880503 2322 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:37.350495 kubelet[2322]: E0515 14:56:37.350461 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:37.352726 kubelet[2322]: E0515 14:56:37.352692 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:37.354436 kubelet[2322]: E0515 14:56:37.354406 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:37.356713 kubelet[2322]: E0515 14:56:37.356684 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:37.358539 kubelet[2322]: E0515 14:56:37.357883 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:37.358936 kubelet[2322]: E0515 14:56:37.358852 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:38.361034 kubelet[2322]: E0515 14:56:38.360992 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:38.363025 kubelet[2322]: E0515 14:56:38.361182 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:38.363025 kubelet[2322]: E0515 14:56:38.361859 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:38.363025 kubelet[2322]: E0515 14:56:38.362013 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:38.363025 kubelet[2322]: E0515 14:56:38.362640 2322 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:38.363025 kubelet[2322]: E0515 14:56:38.362778 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:39.242925 kubelet[2322]: I0515 14:56:39.242802 2322 apiserver.go:52] "Watching apiserver" May 15 14:56:39.255290 kubelet[2322]: E0515 14:56:39.255242 2322 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4334.0.0-a-c1acadba3f\" not found" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.262715 kubelet[2322]: I0515 14:56:39.262675 2322 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 15 14:56:39.340672 kubelet[2322]: I0515 14:56:39.340614 2322 kubelet_node_status.go:79] "Successfully registered node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.359458 kubelet[2322]: I0515 14:56:39.358995 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.364346 kubelet[2322]: I0515 14:56:39.364307 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.365367 kubelet[2322]: I0515 14:56:39.364359 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.378708 kubelet[2322]: E0515 14:56:39.378650 2322 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.379145 kubelet[2322]: E0515 14:56:39.378660 2322 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-c1acadba3f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.379145 kubelet[2322]: I0515 14:56:39.378682 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.379145 kubelet[2322]: E0515 14:56:39.378713 2322 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.379343 kubelet[2322]: E0515 14:56:39.379297 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:39.379560 kubelet[2322]: E0515 14:56:39.379452 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:39.381897 kubelet[2322]: E0515 14:56:39.381860 2322 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.381897 kubelet[2322]: I0515 14:56:39.381890 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:39.384093 kubelet[2322]: E0515 14:56:39.384056 2322 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4334.0.0-a-c1acadba3f\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:40.368266 kubelet[2322]: I0515 14:56:40.368209 2322 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:40.377825 kubelet[2322]: W0515 14:56:40.377705 2322 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:40.378195 kubelet[2322]: E0515 14:56:40.378169 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:41.369047 kubelet[2322]: E0515 14:56:41.368955 2322 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:41.681668 systemd[1]: Reload requested from client PID 2589 ('systemctl') (unit session-7.scope)... May 15 14:56:41.681691 systemd[1]: Reloading... May 15 14:56:41.834559 zram_generator::config[2632]: No configuration found. May 15 14:56:41.990483 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 15 14:56:42.230056 systemd[1]: Reloading finished in 547 ms. May 15 14:56:42.267303 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:42.284286 systemd[1]: kubelet.service: Deactivated successfully. May 15 14:56:42.284680 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:42.284767 systemd[1]: kubelet.service: Consumed 1.019s CPU time, 119.7M memory peak. May 15 14:56:42.288466 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 15 14:56:42.475126 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 15 14:56:42.487201 (kubelet)[2683]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 15 14:56:42.579168 kubelet[2683]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 14:56:42.579168 kubelet[2683]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 15 14:56:42.579168 kubelet[2683]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 15 14:56:42.580216 kubelet[2683]: I0515 14:56:42.578985 2683 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 15 14:56:42.591450 kubelet[2683]: I0515 14:56:42.591393 2683 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 15 14:56:42.592669 kubelet[2683]: I0515 14:56:42.591690 2683 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 15 14:56:42.592669 kubelet[2683]: I0515 14:56:42.592068 2683 server.go:954] "Client rotation is on, will bootstrap in background" May 15 14:56:42.598167 kubelet[2683]: I0515 14:56:42.597133 2683 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 15 14:56:42.602226 kubelet[2683]: I0515 14:56:42.601960 2683 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 15 14:56:42.606951 kubelet[2683]: I0515 14:56:42.606921 2683 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 15 14:56:42.610670 kubelet[2683]: I0515 14:56:42.610629 2683 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 15 14:56:42.610962 kubelet[2683]: I0515 14:56:42.610913 2683 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 15 14:56:42.611226 kubelet[2683]: I0515 14:56:42.610967 2683 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4334.0.0-a-c1acadba3f","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 15 14:56:42.611319 kubelet[2683]: I0515 14:56:42.611236 2683 topology_manager.go:138] "Creating topology manager with none policy" May 15 14:56:42.611319 kubelet[2683]: I0515 14:56:42.611249 2683 container_manager_linux.go:304] "Creating device plugin manager" May 15 14:56:42.611319 kubelet[2683]: I0515 14:56:42.611305 2683 state_mem.go:36] "Initialized new in-memory state store" May 15 14:56:42.611498 kubelet[2683]: I0515 14:56:42.611486 2683 kubelet.go:446] "Attempting to sync node with API server" May 15 14:56:42.611548 kubelet[2683]: I0515 14:56:42.611508 2683 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 15 14:56:42.611584 kubelet[2683]: I0515 14:56:42.611566 2683 kubelet.go:352] "Adding apiserver pod source" May 15 14:56:42.611584 kubelet[2683]: I0515 14:56:42.611577 2683 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 15 14:56:42.617941 kubelet[2683]: I0515 14:56:42.617883 2683 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 15 14:56:42.618421 kubelet[2683]: I0515 14:56:42.618385 2683 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 15 14:56:42.621384 kubelet[2683]: I0515 14:56:42.621348 2683 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 15 14:56:42.621552 kubelet[2683]: I0515 14:56:42.621412 2683 server.go:1287] "Started kubelet" May 15 14:56:42.629085 kubelet[2683]: I0515 14:56:42.629035 2683 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 15 14:56:42.674342 kubelet[2683]: I0515 14:56:42.674272 2683 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 15 14:56:42.676774 kubelet[2683]: I0515 14:56:42.676694 2683 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 15 14:56:42.677222 kubelet[2683]: I0515 14:56:42.677187 2683 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 15 14:56:42.677733 kubelet[2683]: I0515 14:56:42.677660 2683 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 15 14:56:42.681193 kubelet[2683]: I0515 14:56:42.680979 2683 volume_manager.go:297] "Starting Kubelet Volume Manager" May 15 14:56:42.683137 kubelet[2683]: I0515 14:56:42.681950 2683 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 15 14:56:42.683137 kubelet[2683]: I0515 14:56:42.682093 2683 reconciler.go:26] "Reconciler: start to sync state" May 15 14:56:42.683491 kubelet[2683]: I0515 14:56:42.683470 2683 server.go:490] "Adding debug handlers to kubelet server" May 15 14:56:42.686880 kubelet[2683]: E0515 14:56:42.686400 2683 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 15 14:56:42.691584 kubelet[2683]: I0515 14:56:42.690196 2683 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 15 14:56:42.702047 kubelet[2683]: I0515 14:56:42.701984 2683 factory.go:221] Registration of the containerd container factory successfully May 15 14:56:42.702235 kubelet[2683]: I0515 14:56:42.702223 2683 factory.go:221] Registration of the systemd container factory successfully May 15 14:56:42.703673 kubelet[2683]: I0515 14:56:42.703627 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 15 14:56:42.707553 kubelet[2683]: I0515 14:56:42.707441 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 15 14:56:42.709744 kubelet[2683]: I0515 14:56:42.708482 2683 status_manager.go:227] "Starting to sync pod status with apiserver" May 15 14:56:42.709744 kubelet[2683]: I0515 14:56:42.708552 2683 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 15 14:56:42.709744 kubelet[2683]: I0515 14:56:42.708565 2683 kubelet.go:2388] "Starting kubelet main sync loop" May 15 14:56:42.709744 kubelet[2683]: E0515 14:56:42.708644 2683 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800686 2683 cpu_manager.go:221] "Starting CPU manager" policy="none" May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800707 2683 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800734 2683 state_mem.go:36] "Initialized new in-memory state store" May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800938 2683 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800949 2683 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 15 14:56:42.800941 kubelet[2683]: I0515 14:56:42.800972 2683 policy_none.go:49] "None policy: Start" May 15 14:56:42.801374 kubelet[2683]: I0515 14:56:42.800983 2683 memory_manager.go:186] "Starting memorymanager" policy="None" May 15 14:56:42.801374 kubelet[2683]: I0515 14:56:42.800993 2683 state_mem.go:35] "Initializing new in-memory state store" May 15 14:56:42.801374 kubelet[2683]: I0515 14:56:42.801116 2683 state_mem.go:75] "Updated machine memory state" May 15 14:56:42.809175 kubelet[2683]: E0515 14:56:42.809085 2683 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 15 14:56:42.816570 kubelet[2683]: I0515 14:56:42.816320 2683 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 15 14:56:42.817811 kubelet[2683]: I0515 14:56:42.817093 2683 eviction_manager.go:189] "Eviction manager: starting control loop" May 15 14:56:42.817811 kubelet[2683]: I0515 14:56:42.817114 2683 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 15 14:56:42.817811 kubelet[2683]: I0515 14:56:42.817479 2683 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 15 14:56:42.824346 kubelet[2683]: E0515 14:56:42.824052 2683 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 15 14:56:42.930743 kubelet[2683]: I0515 14:56:42.930555 2683 kubelet_node_status.go:76] "Attempting to register node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:42.955375 kubelet[2683]: I0515 14:56:42.955174 2683 kubelet_node_status.go:125] "Node was previously registered" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:42.955375 kubelet[2683]: I0515 14:56:42.955295 2683 kubelet_node_status.go:79] "Successfully registered node" node="ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.011784 kubelet[2683]: I0515 14:56:43.010802 2683 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.014081 kubelet[2683]: I0515 14:56:43.014026 2683 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.016626 kubelet[2683]: I0515 14:56:43.016013 2683 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.030366 kubelet[2683]: W0515 14:56:43.030316 2683 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:43.031264 kubelet[2683]: W0515 14:56:43.031233 2683 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:43.031431 kubelet[2683]: E0515 14:56:43.031318 2683 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" already exists" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.033534 kubelet[2683]: W0515 14:56:43.033478 2683 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:43.086598 kubelet[2683]: I0515 14:56:43.086450 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-k8s-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086598 kubelet[2683]: I0515 14:56:43.086597 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-kubeconfig\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086902 kubelet[2683]: I0515 14:56:43.086630 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-ca-certs\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086902 kubelet[2683]: I0515 14:56:43.086684 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086902 kubelet[2683]: I0515 14:56:43.086739 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-flexvolume-dir\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086902 kubelet[2683]: I0515 14:56:43.086761 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.086902 kubelet[2683]: I0515 14:56:43.086814 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dc64a222dc9b3fd1d547417b64b8c506-kubeconfig\") pod \"kube-scheduler-ci-4334.0.0-a-c1acadba3f\" (UID: \"dc64a222dc9b3fd1d547417b64b8c506\") " pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.087129 kubelet[2683]: I0515 14:56:43.086831 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4b35d03f5ff6f301272d281a76f73df9-k8s-certs\") pod \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" (UID: \"4b35d03f5ff6f301272d281a76f73df9\") " pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.087129 kubelet[2683]: I0515 14:56:43.086846 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3bb4783d2c516c9c05f69ae7d3e1d3cd-ca-certs\") pod \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" (UID: \"3bb4783d2c516c9c05f69ae7d3e1d3cd\") " pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.332510 kubelet[2683]: E0515 14:56:43.332432 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.333156 kubelet[2683]: E0515 14:56:43.333113 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.334704 kubelet[2683]: E0515 14:56:43.334513 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.614104 kubelet[2683]: I0515 14:56:43.613647 2683 apiserver.go:52] "Watching apiserver" May 15 14:56:43.683253 kubelet[2683]: I0515 14:56:43.683130 2683 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 15 14:56:43.751991 kubelet[2683]: E0515 14:56:43.751321 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.751991 kubelet[2683]: I0515 14:56:43.751380 2683 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.751991 kubelet[2683]: I0515 14:56:43.751888 2683 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.790860 kubelet[2683]: W0515 14:56:43.790813 2683 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:43.791036 kubelet[2683]: E0515 14:56:43.790890 2683 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4334.0.0-a-c1acadba3f\" already exists" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.791108 kubelet[2683]: E0515 14:56:43.791085 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.791215 kubelet[2683]: W0515 14:56:43.791193 2683 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 15 14:56:43.791255 kubelet[2683]: E0515 14:56:43.791244 2683 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4334.0.0-a-c1acadba3f\" already exists" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" May 15 14:56:43.791382 kubelet[2683]: E0515 14:56:43.791363 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:43.804542 kubelet[2683]: I0515 14:56:43.804467 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4334.0.0-a-c1acadba3f" podStartSLOduration=3.804439842 podStartE2EDuration="3.804439842s" podCreationTimestamp="2025-05-15 14:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:56:43.804233992 +0000 UTC m=+1.306926892" watchObservedRunningTime="2025-05-15 14:56:43.804439842 +0000 UTC m=+1.307132738" May 15 14:56:43.876804 kubelet[2683]: I0515 14:56:43.876391 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4334.0.0-a-c1acadba3f" podStartSLOduration=0.876367846 podStartE2EDuration="876.367846ms" podCreationTimestamp="2025-05-15 14:56:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:56:43.844438138 +0000 UTC m=+1.347131038" watchObservedRunningTime="2025-05-15 14:56:43.876367846 +0000 UTC m=+1.379060747" May 15 14:56:43.915846 kubelet[2683]: I0515 14:56:43.915122 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4334.0.0-a-c1acadba3f" podStartSLOduration=0.915103586 podStartE2EDuration="915.103586ms" podCreationTimestamp="2025-05-15 14:56:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:56:43.878717409 +0000 UTC m=+1.381410311" watchObservedRunningTime="2025-05-15 14:56:43.915103586 +0000 UTC m=+1.417796486" May 15 14:56:44.756310 kubelet[2683]: E0515 14:56:44.755992 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:44.756310 kubelet[2683]: E0515 14:56:44.756110 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:44.756794 kubelet[2683]: E0515 14:56:44.756562 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:45.761563 kubelet[2683]: E0515 14:56:45.758461 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:46.186284 kubelet[2683]: I0515 14:56:46.186009 2683 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 15 14:56:46.186904 containerd[1559]: time="2025-05-15T14:56:46.186861747Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 15 14:56:46.187399 kubelet[2683]: I0515 14:56:46.187199 2683 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 15 14:56:46.849892 systemd[1]: Created slice kubepods-besteffort-pod9dabdf63_a426_4a22_abf7_899047023d5f.slice - libcontainer container kubepods-besteffort-pod9dabdf63_a426_4a22_abf7_899047023d5f.slice. May 15 14:56:46.922055 kubelet[2683]: I0515 14:56:46.921931 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/9dabdf63-a426-4a22-abf7-899047023d5f-kube-proxy\") pod \"kube-proxy-sgm97\" (UID: \"9dabdf63-a426-4a22-abf7-899047023d5f\") " pod="kube-system/kube-proxy-sgm97" May 15 14:56:46.922055 kubelet[2683]: I0515 14:56:46.921973 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9dabdf63-a426-4a22-abf7-899047023d5f-lib-modules\") pod \"kube-proxy-sgm97\" (UID: \"9dabdf63-a426-4a22-abf7-899047023d5f\") " pod="kube-system/kube-proxy-sgm97" May 15 14:56:46.922055 kubelet[2683]: I0515 14:56:46.921998 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9dabdf63-a426-4a22-abf7-899047023d5f-xtables-lock\") pod \"kube-proxy-sgm97\" (UID: \"9dabdf63-a426-4a22-abf7-899047023d5f\") " pod="kube-system/kube-proxy-sgm97" May 15 14:56:46.922055 kubelet[2683]: I0515 14:56:46.922017 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ggbb\" (UniqueName: \"kubernetes.io/projected/9dabdf63-a426-4a22-abf7-899047023d5f-kube-api-access-9ggbb\") pod \"kube-proxy-sgm97\" (UID: \"9dabdf63-a426-4a22-abf7-899047023d5f\") " pod="kube-system/kube-proxy-sgm97" May 15 14:56:47.161970 kubelet[2683]: E0515 14:56:47.160968 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:47.163707 containerd[1559]: time="2025-05-15T14:56:47.163643672Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sgm97,Uid:9dabdf63-a426-4a22-abf7-899047023d5f,Namespace:kube-system,Attempt:0,}" May 15 14:56:47.189307 containerd[1559]: time="2025-05-15T14:56:47.189164932Z" level=info msg="connecting to shim 5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb" address="unix:///run/containerd/s/d3894580ab8eff05be8a28f82cd557d95ca9172a1e32efa238f3db26ae28f5ef" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:47.247853 systemd[1]: Started cri-containerd-5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb.scope - libcontainer container 5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb. May 15 14:56:47.254337 kubelet[2683]: E0515 14:56:47.254102 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:47.316787 containerd[1559]: time="2025-05-15T14:56:47.316702207Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-sgm97,Uid:9dabdf63-a426-4a22-abf7-899047023d5f,Namespace:kube-system,Attempt:0,} returns sandbox id \"5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb\"" May 15 14:56:47.318941 kubelet[2683]: E0515 14:56:47.318907 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:47.323462 containerd[1559]: time="2025-05-15T14:56:47.323292721Z" level=info msg="CreateContainer within sandbox \"5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 15 14:56:47.347558 containerd[1559]: time="2025-05-15T14:56:47.347403459Z" level=info msg="Container 548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:47.355426 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3610565872.mount: Deactivated successfully. May 15 14:56:47.375839 systemd[1]: Created slice kubepods-besteffort-pod32afc8d0_2678_48ff_97a2_c7d5de52688d.slice - libcontainer container kubepods-besteffort-pod32afc8d0_2678_48ff_97a2_c7d5de52688d.slice. May 15 14:56:47.384552 containerd[1559]: time="2025-05-15T14:56:47.383274304Z" level=info msg="CreateContainer within sandbox \"5b6efb66a3e2a1f579f5098621c6df26f5ea12c3bf02912dec53e45720a65cfb\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5\"" May 15 14:56:47.386006 containerd[1559]: time="2025-05-15T14:56:47.385897598Z" level=info msg="StartContainer for \"548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5\"" May 15 14:56:47.387931 containerd[1559]: time="2025-05-15T14:56:47.387885954Z" level=info msg="connecting to shim 548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5" address="unix:///run/containerd/s/d3894580ab8eff05be8a28f82cd557d95ca9172a1e32efa238f3db26ae28f5ef" protocol=ttrpc version=3 May 15 14:56:47.426412 kubelet[2683]: I0515 14:56:47.425800 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p98xx\" (UniqueName: \"kubernetes.io/projected/32afc8d0-2678-48ff-97a2-c7d5de52688d-kube-api-access-p98xx\") pod \"tigera-operator-789496d6f5-m6ldv\" (UID: \"32afc8d0-2678-48ff-97a2-c7d5de52688d\") " pod="tigera-operator/tigera-operator-789496d6f5-m6ldv" May 15 14:56:47.426412 kubelet[2683]: I0515 14:56:47.425845 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/32afc8d0-2678-48ff-97a2-c7d5de52688d-var-lib-calico\") pod \"tigera-operator-789496d6f5-m6ldv\" (UID: \"32afc8d0-2678-48ff-97a2-c7d5de52688d\") " pod="tigera-operator/tigera-operator-789496d6f5-m6ldv" May 15 14:56:47.427045 systemd[1]: Started cri-containerd-548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5.scope - libcontainer container 548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5. May 15 14:56:47.488893 containerd[1559]: time="2025-05-15T14:56:47.488821069Z" level=info msg="StartContainer for \"548247e0bbc421bd217e4e0bcde6803a5dc5a88b18b516e948cc102b4cde3dc5\" returns successfully" May 15 14:56:47.685584 containerd[1559]: time="2025-05-15T14:56:47.684818195Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-m6ldv,Uid:32afc8d0-2678-48ff-97a2-c7d5de52688d,Namespace:tigera-operator,Attempt:0,}" May 15 14:56:47.713630 containerd[1559]: time="2025-05-15T14:56:47.713473401Z" level=info msg="connecting to shim 54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0" address="unix:///run/containerd/s/c2630bb41bce502750dc60ece447ebb6886d17e6b2a6b3fd6f1cd81b8009c1bc" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:47.748806 systemd[1]: Started cri-containerd-54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0.scope - libcontainer container 54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0. May 15 14:56:47.768587 kubelet[2683]: E0515 14:56:47.768312 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:47.768587 kubelet[2683]: E0515 14:56:47.768315 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:47.814257 kubelet[2683]: I0515 14:56:47.813891 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-sgm97" podStartSLOduration=1.813869973 podStartE2EDuration="1.813869973s" podCreationTimestamp="2025-05-15 14:56:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:56:47.812115278 +0000 UTC m=+5.314808180" watchObservedRunningTime="2025-05-15 14:56:47.813869973 +0000 UTC m=+5.316562874" May 15 14:56:47.867866 containerd[1559]: time="2025-05-15T14:56:47.867729995Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-m6ldv,Uid:32afc8d0-2678-48ff-97a2-c7d5de52688d,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0\"" May 15 14:56:47.871652 containerd[1559]: time="2025-05-15T14:56:47.871479959Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 15 14:56:47.875704 systemd-resolved[1401]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. May 15 14:56:48.640353 sudo[1772]: pam_unix(sudo:session): session closed for user root May 15 14:56:48.644076 sshd[1771]: Connection closed by 139.178.68.195 port 50840 May 15 14:56:48.646869 sshd-session[1769]: pam_unix(sshd:session): session closed for user core May 15 14:56:48.652178 systemd[1]: sshd@6-24.199.109.202:22-139.178.68.195:50840.service: Deactivated successfully. May 15 14:56:48.656188 systemd[1]: session-7.scope: Deactivated successfully. May 15 14:56:48.656633 systemd[1]: session-7.scope: Consumed 6.223s CPU time, 162.1M memory peak. May 15 14:56:48.659891 systemd-logind[1529]: Session 7 logged out. Waiting for processes to exit. May 15 14:56:48.662462 systemd-logind[1529]: Removed session 7. May 15 14:56:49.242160 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount920725479.mount: Deactivated successfully. May 15 14:56:50.002197 containerd[1559]: time="2025-05-15T14:56:50.002103564Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:50.003939 containerd[1559]: time="2025-05-15T14:56:50.003863925Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=22002662" May 15 14:56:50.005670 containerd[1559]: time="2025-05-15T14:56:50.005598596Z" level=info msg="ImageCreate event name:\"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:50.007805 containerd[1559]: time="2025-05-15T14:56:50.007728947Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:50.009687 containerd[1559]: time="2025-05-15T14:56:50.009594970Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"21998657\" in 2.138068267s" May 15 14:56:50.009687 containerd[1559]: time="2025-05-15T14:56:50.009651781Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" May 15 14:56:50.013107 containerd[1559]: time="2025-05-15T14:56:50.013015731Z" level=info msg="CreateContainer within sandbox \"54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 15 14:56:50.021739 containerd[1559]: time="2025-05-15T14:56:50.021682748Z" level=info msg="Container 2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:50.041546 containerd[1559]: time="2025-05-15T14:56:50.041011472Z" level=info msg="CreateContainer within sandbox \"54237018a59d8eb9fc6dd5af00f882f660e4a0dd6ad6d11887dec308c47c01c0\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c\"" May 15 14:56:50.046701 containerd[1559]: time="2025-05-15T14:56:50.046407662Z" level=info msg="StartContainer for \"2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c\"" May 15 14:56:50.050182 containerd[1559]: time="2025-05-15T14:56:50.050135104Z" level=info msg="connecting to shim 2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c" address="unix:///run/containerd/s/c2630bb41bce502750dc60ece447ebb6886d17e6b2a6b3fd6f1cd81b8009c1bc" protocol=ttrpc version=3 May 15 14:56:50.090853 systemd[1]: Started cri-containerd-2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c.scope - libcontainer container 2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c. May 15 14:56:50.140236 containerd[1559]: time="2025-05-15T14:56:50.139559753Z" level=info msg="StartContainer for \"2c5c418436294978552b9581bc3cd2e5a96761fbec24637477ae4fe52f7b474c\" returns successfully" May 15 14:56:51.510922 kubelet[2683]: E0515 14:56:51.510874 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:51.538995 kubelet[2683]: I0515 14:56:51.538932 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-789496d6f5-m6ldv" podStartSLOduration=2.3982573289999998 podStartE2EDuration="4.538909846s" podCreationTimestamp="2025-05-15 14:56:47 +0000 UTC" firstStartedPulling="2025-05-15 14:56:47.870414464 +0000 UTC m=+5.373107356" lastFinishedPulling="2025-05-15 14:56:50.011066977 +0000 UTC m=+7.513759873" observedRunningTime="2025-05-15 14:56:50.792147525 +0000 UTC m=+8.294840426" watchObservedRunningTime="2025-05-15 14:56:51.538909846 +0000 UTC m=+9.041602746" May 15 14:56:51.781228 kubelet[2683]: E0515 14:56:51.780845 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:53.627693 systemd[1]: Created slice kubepods-besteffort-podfd270830_c56c_4d9f_8dd7_8181a45e3381.slice - libcontainer container kubepods-besteffort-podfd270830_c56c_4d9f_8dd7_8181a45e3381.slice. May 15 14:56:53.670557 kubelet[2683]: I0515 14:56:53.670056 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd270830-c56c-4d9f-8dd7-8181a45e3381-tigera-ca-bundle\") pod \"calico-typha-6f9c8647b9-b52kj\" (UID: \"fd270830-c56c-4d9f-8dd7-8181a45e3381\") " pod="calico-system/calico-typha-6f9c8647b9-b52kj" May 15 14:56:53.671362 kubelet[2683]: I0515 14:56:53.671212 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5rz\" (UniqueName: \"kubernetes.io/projected/fd270830-c56c-4d9f-8dd7-8181a45e3381-kube-api-access-nc5rz\") pod \"calico-typha-6f9c8647b9-b52kj\" (UID: \"fd270830-c56c-4d9f-8dd7-8181a45e3381\") " pod="calico-system/calico-typha-6f9c8647b9-b52kj" May 15 14:56:53.671362 kubelet[2683]: I0515 14:56:53.671289 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/fd270830-c56c-4d9f-8dd7-8181a45e3381-typha-certs\") pod \"calico-typha-6f9c8647b9-b52kj\" (UID: \"fd270830-c56c-4d9f-8dd7-8181a45e3381\") " pod="calico-system/calico-typha-6f9c8647b9-b52kj" May 15 14:56:53.770948 systemd[1]: Created slice kubepods-besteffort-pod5cec4f76_4291_4626_9ede_35d15f7f59c4.slice - libcontainer container kubepods-besteffort-pod5cec4f76_4291_4626_9ede_35d15f7f59c4.slice. May 15 14:56:53.873733 kubelet[2683]: I0515 14:56:53.873588 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-var-lib-calico\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.874765 kubelet[2683]: I0515 14:56:53.874674 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-cni-net-dir\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875018 kubelet[2683]: I0515 14:56:53.874996 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-cni-log-dir\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875215 kubelet[2683]: I0515 14:56:53.875194 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/5cec4f76-4291-4626-9ede-35d15f7f59c4-node-certs\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875457 kubelet[2683]: I0515 14:56:53.875398 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh25v\" (UniqueName: \"kubernetes.io/projected/5cec4f76-4291-4626-9ede-35d15f7f59c4-kube-api-access-fh25v\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875991 kubelet[2683]: I0515 14:56:53.875705 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-lib-modules\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875991 kubelet[2683]: I0515 14:56:53.875744 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-cni-bin-dir\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875991 kubelet[2683]: I0515 14:56:53.875764 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-policysync\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875991 kubelet[2683]: I0515 14:56:53.875785 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-xtables-lock\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.875991 kubelet[2683]: I0515 14:56:53.875808 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-flexvol-driver-host\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.876255 kubelet[2683]: I0515 14:56:53.875834 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cec4f76-4291-4626-9ede-35d15f7f59c4-tigera-ca-bundle\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.876255 kubelet[2683]: I0515 14:56:53.875876 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/5cec4f76-4291-4626-9ede-35d15f7f59c4-var-run-calico\") pod \"calico-node-7c27f\" (UID: \"5cec4f76-4291-4626-9ede-35d15f7f59c4\") " pod="calico-system/calico-node-7c27f" May 15 14:56:53.905350 kubelet[2683]: E0515 14:56:53.905056 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:56:53.935809 kubelet[2683]: E0515 14:56:53.935661 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:53.937840 containerd[1559]: time="2025-05-15T14:56:53.937794463Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6f9c8647b9-b52kj,Uid:fd270830-c56c-4d9f-8dd7-8181a45e3381,Namespace:calico-system,Attempt:0,}" May 15 14:56:53.976282 containerd[1559]: time="2025-05-15T14:56:53.976172291Z" level=info msg="connecting to shim 4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b" address="unix:///run/containerd/s/b1941b435573fa90d98c155f326202f89cc8e3eabf626f270c61ae966554be22" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:53.977647 kubelet[2683]: I0515 14:56:53.977162 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3300dd9b-523b-416e-9de8-89b5d9c22fa3-socket-dir\") pod \"csi-node-driver-5pnkc\" (UID: \"3300dd9b-523b-416e-9de8-89b5d9c22fa3\") " pod="calico-system/csi-node-driver-5pnkc" May 15 14:56:53.977647 kubelet[2683]: I0515 14:56:53.977213 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46nqx\" (UniqueName: \"kubernetes.io/projected/3300dd9b-523b-416e-9de8-89b5d9c22fa3-kube-api-access-46nqx\") pod \"csi-node-driver-5pnkc\" (UID: \"3300dd9b-523b-416e-9de8-89b5d9c22fa3\") " pod="calico-system/csi-node-driver-5pnkc" May 15 14:56:53.977647 kubelet[2683]: I0515 14:56:53.977256 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/3300dd9b-523b-416e-9de8-89b5d9c22fa3-varrun\") pod \"csi-node-driver-5pnkc\" (UID: \"3300dd9b-523b-416e-9de8-89b5d9c22fa3\") " pod="calico-system/csi-node-driver-5pnkc" May 15 14:56:53.977647 kubelet[2683]: I0515 14:56:53.977404 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3300dd9b-523b-416e-9de8-89b5d9c22fa3-registration-dir\") pod \"csi-node-driver-5pnkc\" (UID: \"3300dd9b-523b-416e-9de8-89b5d9c22fa3\") " pod="calico-system/csi-node-driver-5pnkc" May 15 14:56:53.977647 kubelet[2683]: I0515 14:56:53.977458 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3300dd9b-523b-416e-9de8-89b5d9c22fa3-kubelet-dir\") pod \"csi-node-driver-5pnkc\" (UID: \"3300dd9b-523b-416e-9de8-89b5d9c22fa3\") " pod="calico-system/csi-node-driver-5pnkc" May 15 14:56:53.983826 kubelet[2683]: E0515 14:56:53.983790 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.984543 kubelet[2683]: W0515 14:56:53.984436 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.984875 kubelet[2683]: E0515 14:56:53.984510 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.985643 kubelet[2683]: E0515 14:56:53.985605 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.985879 kubelet[2683]: W0515 14:56:53.985726 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.985879 kubelet[2683]: E0515 14:56:53.985765 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.987020 kubelet[2683]: E0515 14:56:53.986741 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.987020 kubelet[2683]: W0515 14:56:53.986758 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.987271 kubelet[2683]: E0515 14:56:53.987209 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.988601 kubelet[2683]: E0515 14:56:53.988579 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.988601 kubelet[2683]: W0515 14:56:53.988596 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.989349 kubelet[2683]: E0515 14:56:53.988619 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.989408 kubelet[2683]: E0515 14:56:53.989365 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.989920 kubelet[2683]: W0515 14:56:53.989378 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.990066 kubelet[2683]: E0515 14:56:53.989964 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.991372 kubelet[2683]: E0515 14:56:53.991350 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.991372 kubelet[2683]: W0515 14:56:53.991367 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.991767 kubelet[2683]: E0515 14:56:53.991604 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.993718 kubelet[2683]: E0515 14:56:53.993652 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.993718 kubelet[2683]: W0515 14:56:53.993701 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.993718 kubelet[2683]: E0515 14:56:53.993751 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.994489 kubelet[2683]: E0515 14:56:53.994465 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.994920 kubelet[2683]: W0515 14:56:53.994774 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.995125 kubelet[2683]: E0515 14:56:53.994815 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.995541 kubelet[2683]: E0515 14:56:53.995434 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.995541 kubelet[2683]: W0515 14:56:53.995452 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.995541 kubelet[2683]: E0515 14:56:53.995478 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.997343 kubelet[2683]: E0515 14:56:53.997318 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.997343 kubelet[2683]: W0515 14:56:53.997338 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:53.997538 kubelet[2683]: E0515 14:56:53.997372 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:53.999570 kubelet[2683]: E0515 14:56:53.998800 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:53.999678 kubelet[2683]: W0515 14:56:53.999575 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.000122 kubelet[2683]: E0515 14:56:53.999739 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.000410 kubelet[2683]: E0515 14:56:54.000388 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.000410 kubelet[2683]: W0515 14:56:54.000407 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.000537 kubelet[2683]: E0515 14:56:54.000429 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.021656 kubelet[2683]: E0515 14:56:54.021596 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.022611 kubelet[2683]: W0515 14:56:54.021767 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.022611 kubelet[2683]: E0515 14:56:54.021798 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.056163 systemd[1]: Started cri-containerd-4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b.scope - libcontainer container 4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b. May 15 14:56:54.061126 kubelet[2683]: E0515 14:56:54.061096 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.061126 kubelet[2683]: W0515 14:56:54.061133 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.061610 kubelet[2683]: E0515 14:56:54.061156 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.080271 kubelet[2683]: E0515 14:56:54.079659 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.080271 kubelet[2683]: W0515 14:56:54.079689 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.080271 kubelet[2683]: E0515 14:56:54.079713 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.080271 kubelet[2683]: E0515 14:56:54.079794 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:54.080271 kubelet[2683]: E0515 14:56:54.079972 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.080271 kubelet[2683]: W0515 14:56:54.079982 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.080271 kubelet[2683]: E0515 14:56:54.079994 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.080723 kubelet[2683]: E0515 14:56:54.080622 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.080723 kubelet[2683]: W0515 14:56:54.080638 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.080723 kubelet[2683]: E0515 14:56:54.080652 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.082628 kubelet[2683]: E0515 14:56:54.082597 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.082628 kubelet[2683]: W0515 14:56:54.082620 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.082628 kubelet[2683]: E0515 14:56:54.082647 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.083054 containerd[1559]: time="2025-05-15T14:56:54.082483424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-7c27f,Uid:5cec4f76-4291-4626-9ede-35d15f7f59c4,Namespace:calico-system,Attempt:0,}" May 15 14:56:54.084109 kubelet[2683]: E0515 14:56:54.084082 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.084383 kubelet[2683]: W0515 14:56:54.084228 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.084383 kubelet[2683]: E0515 14:56:54.084260 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.085354 kubelet[2683]: E0515 14:56:54.085236 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.085354 kubelet[2683]: W0515 14:56:54.085253 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.085354 kubelet[2683]: E0515 14:56:54.085274 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.086130 kubelet[2683]: E0515 14:56:54.086110 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.086130 kubelet[2683]: W0515 14:56:54.086126 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.086623 kubelet[2683]: E0515 14:56:54.086586 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.087975 kubelet[2683]: E0515 14:56:54.087252 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.087975 kubelet[2683]: W0515 14:56:54.087836 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.087975 kubelet[2683]: E0515 14:56:54.087875 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.088443 kubelet[2683]: E0515 14:56:54.088366 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.088443 kubelet[2683]: W0515 14:56:54.088381 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.088443 kubelet[2683]: E0515 14:56:54.088414 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.089109 kubelet[2683]: E0515 14:56:54.089025 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.089109 kubelet[2683]: W0515 14:56:54.089044 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.089776 kubelet[2683]: E0515 14:56:54.089097 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.090111 kubelet[2683]: E0515 14:56:54.090058 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.090111 kubelet[2683]: W0515 14:56:54.090076 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.090219 kubelet[2683]: E0515 14:56:54.090178 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.091080 kubelet[2683]: E0515 14:56:54.090999 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.091080 kubelet[2683]: W0515 14:56:54.091017 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.091396 kubelet[2683]: E0515 14:56:54.091244 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.091925 kubelet[2683]: E0515 14:56:54.091813 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.091925 kubelet[2683]: W0515 14:56:54.091830 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.092764 kubelet[2683]: E0515 14:56:54.092116 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.093555 kubelet[2683]: E0515 14:56:54.093189 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.093555 kubelet[2683]: W0515 14:56:54.093208 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.093833 kubelet[2683]: E0515 14:56:54.093739 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.094262 kubelet[2683]: E0515 14:56:54.094178 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.094262 kubelet[2683]: W0515 14:56:54.094197 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.094262 kubelet[2683]: E0515 14:56:54.094240 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.095428 kubelet[2683]: E0515 14:56:54.095254 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.095428 kubelet[2683]: W0515 14:56:54.095272 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.095428 kubelet[2683]: E0515 14:56:54.095335 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.096584 kubelet[2683]: E0515 14:56:54.096562 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.096803 kubelet[2683]: W0515 14:56:54.096688 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.097054 kubelet[2683]: E0515 14:56:54.096958 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.097330 kubelet[2683]: E0515 14:56:54.097235 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.097330 kubelet[2683]: W0515 14:56:54.097251 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.097330 kubelet[2683]: E0515 14:56:54.097292 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.099278 kubelet[2683]: E0515 14:56:54.099132 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.099278 kubelet[2683]: W0515 14:56:54.099153 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.099278 kubelet[2683]: E0515 14:56:54.099205 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.099912 kubelet[2683]: E0515 14:56:54.099720 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.099912 kubelet[2683]: W0515 14:56:54.099737 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.101093 kubelet[2683]: E0515 14:56:54.100630 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.101093 kubelet[2683]: E0515 14:56:54.100769 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.101093 kubelet[2683]: W0515 14:56:54.100781 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.101093 kubelet[2683]: E0515 14:56:54.100805 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.102170 kubelet[2683]: E0515 14:56:54.101729 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.102340 kubelet[2683]: W0515 14:56:54.102260 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.102589 kubelet[2683]: E0515 14:56:54.102574 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.102781 kubelet[2683]: W0515 14:56:54.102656 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.103275 kubelet[2683]: E0515 14:56:54.103257 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.103394 kubelet[2683]: W0515 14:56:54.103378 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.103472 kubelet[2683]: E0515 14:56:54.103458 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.104638 kubelet[2683]: E0515 14:56:54.103610 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.104638 kubelet[2683]: E0515 14:56:54.103629 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.105203 kubelet[2683]: E0515 14:56:54.104977 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.105203 kubelet[2683]: W0515 14:56:54.104995 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.105203 kubelet[2683]: E0515 14:56:54.105012 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.119615 containerd[1559]: time="2025-05-15T14:56:54.119507808Z" level=info msg="connecting to shim 4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1" address="unix:///run/containerd/s/171f23b7b14ba40beb4a1e82dcc0e417284b841849914f11bd47021ec53e26f0" namespace=k8s.io protocol=ttrpc version=3 May 15 14:56:54.132594 kubelet[2683]: E0515 14:56:54.132516 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.132811 kubelet[2683]: W0515 14:56:54.132790 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.133267 kubelet[2683]: E0515 14:56:54.132932 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.171084 systemd[1]: Started cri-containerd-4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1.scope - libcontainer container 4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1. May 15 14:56:54.248221 containerd[1559]: time="2025-05-15T14:56:54.248161366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-7c27f,Uid:5cec4f76-4291-4626-9ede-35d15f7f59c4,Namespace:calico-system,Attempt:0,} returns sandbox id \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\"" May 15 14:56:54.249684 kubelet[2683]: E0515 14:56:54.249561 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:54.253561 containerd[1559]: time="2025-05-15T14:56:54.253480496Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 15 14:56:54.271312 containerd[1559]: time="2025-05-15T14:56:54.270755177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6f9c8647b9-b52kj,Uid:fd270830-c56c-4d9f-8dd7-8181a45e3381,Namespace:calico-system,Attempt:0,} returns sandbox id \"4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b\"" May 15 14:56:54.273801 kubelet[2683]: E0515 14:56:54.273742 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:54.903075 kubelet[2683]: E0515 14:56:54.903035 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:54.962573 kubelet[2683]: E0515 14:56:54.962410 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.962573 kubelet[2683]: W0515 14:56:54.962438 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.962573 kubelet[2683]: E0515 14:56:54.962462 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.962831 kubelet[2683]: E0515 14:56:54.962679 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.962831 kubelet[2683]: W0515 14:56:54.962694 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.962831 kubelet[2683]: E0515 14:56:54.962705 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.962996 kubelet[2683]: E0515 14:56:54.962857 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.962996 kubelet[2683]: W0515 14:56:54.962864 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.962996 kubelet[2683]: E0515 14:56:54.962873 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.963218 kubelet[2683]: E0515 14:56:54.963054 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.963218 kubelet[2683]: W0515 14:56:54.963063 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.963218 kubelet[2683]: E0515 14:56:54.963074 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:54.963355 kubelet[2683]: E0515 14:56:54.963276 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:54.963355 kubelet[2683]: W0515 14:56:54.963284 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:54.963355 kubelet[2683]: E0515 14:56:54.963293 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:55.709775 kubelet[2683]: E0515 14:56:55.709701 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:56:55.810862 kubelet[2683]: E0515 14:56:55.810805 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:55.868848 kubelet[2683]: E0515 14:56:55.868612 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:55.868848 kubelet[2683]: W0515 14:56:55.868656 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:55.868848 kubelet[2683]: E0515 14:56:55.868689 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:55.869418 kubelet[2683]: E0515 14:56:55.869152 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:55.869418 kubelet[2683]: W0515 14:56:55.869170 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:55.869418 kubelet[2683]: E0515 14:56:55.869189 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:55.869674 kubelet[2683]: E0515 14:56:55.869656 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:55.869911 kubelet[2683]: W0515 14:56:55.869757 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:55.869911 kubelet[2683]: E0515 14:56:55.869776 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:55.870094 kubelet[2683]: E0515 14:56:55.870079 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:55.870176 kubelet[2683]: W0515 14:56:55.870150 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:55.870252 kubelet[2683]: E0515 14:56:55.870238 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:55.870723 kubelet[2683]: E0515 14:56:55.870607 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 15 14:56:55.870723 kubelet[2683]: W0515 14:56:55.870623 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 15 14:56:55.870723 kubelet[2683]: E0515 14:56:55.870637 2683 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 15 14:56:56.114775 containerd[1559]: time="2025-05-15T14:56:56.114603998Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:56.117034 containerd[1559]: time="2025-05-15T14:56:56.116826958Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5366937" May 15 14:56:56.120757 containerd[1559]: time="2025-05-15T14:56:56.120680493Z" level=info msg="ImageCreate event name:\"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:56.127727 containerd[1559]: time="2025-05-15T14:56:56.127503218Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:56.130079 containerd[1559]: time="2025-05-15T14:56:56.130029544Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6859519\" in 1.876069333s" May 15 14:56:56.130785 containerd[1559]: time="2025-05-15T14:56:56.130595276Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" May 15 14:56:56.133102 containerd[1559]: time="2025-05-15T14:56:56.132788530Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 15 14:56:56.138049 containerd[1559]: time="2025-05-15T14:56:56.137565681Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 15 14:56:56.149058 containerd[1559]: time="2025-05-15T14:56:56.148992909Z" level=info msg="Container a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:56.153423 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount394094327.mount: Deactivated successfully. May 15 14:56:56.168549 containerd[1559]: time="2025-05-15T14:56:56.168480338Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\"" May 15 14:56:56.171911 containerd[1559]: time="2025-05-15T14:56:56.171837284Z" level=info msg="StartContainer for \"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\"" May 15 14:56:56.176828 containerd[1559]: time="2025-05-15T14:56:56.176600644Z" level=info msg="connecting to shim a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176" address="unix:///run/containerd/s/171f23b7b14ba40beb4a1e82dcc0e417284b841849914f11bd47021ec53e26f0" protocol=ttrpc version=3 May 15 14:56:56.209813 systemd[1]: Started cri-containerd-a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176.scope - libcontainer container a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176. May 15 14:56:56.272741 containerd[1559]: time="2025-05-15T14:56:56.272691595Z" level=info msg="StartContainer for \"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\" returns successfully" May 15 14:56:56.288244 systemd[1]: cri-containerd-a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176.scope: Deactivated successfully. May 15 14:56:56.294001 containerd[1559]: time="2025-05-15T14:56:56.293940738Z" level=info msg="received exit event container_id:\"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\" id:\"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\" pid:3235 exited_at:{seconds:1747321016 nanos:291716956}" May 15 14:56:56.294542 containerd[1559]: time="2025-05-15T14:56:56.294029958Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\" id:\"a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176\" pid:3235 exited_at:{seconds:1747321016 nanos:291716956}" May 15 14:56:56.330198 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a411fc16d4ec90121b0380fe5fec2c92bf4405d7f71e7842061d1afef5a2d176-rootfs.mount: Deactivated successfully. May 15 14:56:56.518249 update_engine[1531]: I20250515 14:56:56.517783 1531 update_attempter.cc:509] Updating boot flags... May 15 14:56:56.835627 kubelet[2683]: E0515 14:56:56.833038 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:56:57.710065 kubelet[2683]: E0515 14:56:57.709999 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:56:58.839549 containerd[1559]: time="2025-05-15T14:56:58.839331598Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:58.841538 containerd[1559]: time="2025-05-15T14:56:58.841461599Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=30426870" May 15 14:56:58.841857 containerd[1559]: time="2025-05-15T14:56:58.841836982Z" level=info msg="ImageCreate event name:\"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:58.844555 containerd[1559]: time="2025-05-15T14:56:58.844498530Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:56:58.846345 containerd[1559]: time="2025-05-15T14:56:58.846298725Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"31919484\" in 2.712531919s" May 15 14:56:58.846345 containerd[1559]: time="2025-05-15T14:56:58.846348814Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" May 15 14:56:58.851589 containerd[1559]: time="2025-05-15T14:56:58.851540748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 15 14:56:58.870807 containerd[1559]: time="2025-05-15T14:56:58.870710160Z" level=info msg="CreateContainer within sandbox \"4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 15 14:56:58.881246 containerd[1559]: time="2025-05-15T14:56:58.880475101Z" level=info msg="Container 0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481: CDI devices from CRI Config.CDIDevices: []" May 15 14:56:58.889954 containerd[1559]: time="2025-05-15T14:56:58.889903378Z" level=info msg="CreateContainer within sandbox \"4a0db7ebeac13f8da45615e358fd604d3ad8246d5e47323e1ac1408d32f3bd4b\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481\"" May 15 14:56:58.891844 containerd[1559]: time="2025-05-15T14:56:58.891805270Z" level=info msg="StartContainer for \"0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481\"" May 15 14:56:58.895471 containerd[1559]: time="2025-05-15T14:56:58.895329879Z" level=info msg="connecting to shim 0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481" address="unix:///run/containerd/s/b1941b435573fa90d98c155f326202f89cc8e3eabf626f270c61ae966554be22" protocol=ttrpc version=3 May 15 14:56:58.933803 systemd[1]: Started cri-containerd-0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481.scope - libcontainer container 0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481. May 15 14:56:59.006688 containerd[1559]: time="2025-05-15T14:56:59.006607466Z" level=info msg="StartContainer for \"0a92d0fba462f5f5fbb61370a27bded261aaea4708c1106a7de5a205b36ef481\" returns successfully" May 15 14:56:59.710782 kubelet[2683]: E0515 14:56:59.710670 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:56:59.847154 kubelet[2683]: E0515 14:56:59.847110 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:00.854188 kubelet[2683]: I0515 14:57:00.853874 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 14:57:00.855148 kubelet[2683]: E0515 14:57:00.855102 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:01.710282 kubelet[2683]: E0515 14:57:01.709741 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:57:03.709594 kubelet[2683]: E0515 14:57:03.709542 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:57:04.744396 containerd[1559]: time="2025-05-15T14:57:04.744240344Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:04.745586 containerd[1559]: time="2025-05-15T14:57:04.745492178Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=97793683" May 15 14:57:04.750279 containerd[1559]: time="2025-05-15T14:57:04.749339634Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"99286305\" in 5.897744813s" May 15 14:57:04.750279 containerd[1559]: time="2025-05-15T14:57:04.749397910Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" May 15 14:57:04.753048 containerd[1559]: time="2025-05-15T14:57:04.751046728Z" level=info msg="ImageCreate event name:\"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:04.753048 containerd[1559]: time="2025-05-15T14:57:04.752207299Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:04.755882 containerd[1559]: time="2025-05-15T14:57:04.755827770Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 15 14:57:04.768788 containerd[1559]: time="2025-05-15T14:57:04.768722645Z" level=info msg="Container 70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:04.781223 containerd[1559]: time="2025-05-15T14:57:04.781166815Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\"" May 15 14:57:04.782376 containerd[1559]: time="2025-05-15T14:57:04.782289675Z" level=info msg="StartContainer for \"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\"" May 15 14:57:04.785510 containerd[1559]: time="2025-05-15T14:57:04.785409586Z" level=info msg="connecting to shim 70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8" address="unix:///run/containerd/s/171f23b7b14ba40beb4a1e82dcc0e417284b841849914f11bd47021ec53e26f0" protocol=ttrpc version=3 May 15 14:57:04.823778 systemd[1]: Started cri-containerd-70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8.scope - libcontainer container 70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8. May 15 14:57:04.890341 containerd[1559]: time="2025-05-15T14:57:04.890177679Z" level=info msg="StartContainer for \"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\" returns successfully" May 15 14:57:05.479388 systemd[1]: cri-containerd-70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8.scope: Deactivated successfully. May 15 14:57:05.479882 systemd[1]: cri-containerd-70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8.scope: Consumed 643ms CPU time, 149.7M memory peak, 708K read from disk, 154M written to disk. May 15 14:57:05.509317 containerd[1559]: time="2025-05-15T14:57:05.509229515Z" level=info msg="TaskExit event in podsandbox handler container_id:\"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\" id:\"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\" pid:3351 exited_at:{seconds:1747321025 nanos:482701514}" May 15 14:57:05.517787 containerd[1559]: time="2025-05-15T14:57:05.517680084Z" level=info msg="received exit event container_id:\"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\" id:\"70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8\" pid:3351 exited_at:{seconds:1747321025 nanos:482701514}" May 15 14:57:05.555567 kubelet[2683]: I0515 14:57:05.554580 2683 kubelet_node_status.go:502] "Fast updating node status as it just became ready" May 15 14:57:05.579011 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-70c1013bcb4375642a580053d42689f51a11e25dd44b96c6e964c977492b33b8-rootfs.mount: Deactivated successfully. May 15 14:57:05.658906 kubelet[2683]: I0515 14:57:05.658817 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6f9c8647b9-b52kj" podStartSLOduration=8.087333636 podStartE2EDuration="12.658791499s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:56:54.276231882 +0000 UTC m=+11.778924780" lastFinishedPulling="2025-05-15 14:56:58.847689744 +0000 UTC m=+16.350382643" observedRunningTime="2025-05-15 14:56:59.868970828 +0000 UTC m=+17.371663729" watchObservedRunningTime="2025-05-15 14:57:05.658791499 +0000 UTC m=+23.161484448" May 15 14:57:05.676193 systemd[1]: Created slice kubepods-burstable-poded3eb677_7d86_464b_9053_ccf595c17dc1.slice - libcontainer container kubepods-burstable-poded3eb677_7d86_464b_9053_ccf595c17dc1.slice. May 15 14:57:05.681879 kubelet[2683]: I0515 14:57:05.681812 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnvv9\" (UniqueName: \"kubernetes.io/projected/ed3eb677-7d86-464b-9053-ccf595c17dc1-kube-api-access-pnvv9\") pod \"coredns-668d6bf9bc-wbvvg\" (UID: \"ed3eb677-7d86-464b-9053-ccf595c17dc1\") " pod="kube-system/coredns-668d6bf9bc-wbvvg" May 15 14:57:05.681879 kubelet[2683]: I0515 14:57:05.681853 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed3eb677-7d86-464b-9053-ccf595c17dc1-config-volume\") pod \"coredns-668d6bf9bc-wbvvg\" (UID: \"ed3eb677-7d86-464b-9053-ccf595c17dc1\") " pod="kube-system/coredns-668d6bf9bc-wbvvg" May 15 14:57:05.700255 systemd[1]: Created slice kubepods-burstable-pode458056c_828c_4a75_acfb_1672e71f964a.slice - libcontainer container kubepods-burstable-pode458056c_828c_4a75_acfb_1672e71f964a.slice. May 15 14:57:05.716676 systemd[1]: Created slice kubepods-besteffort-pod7d0ff7f0_5822_4b30_acc7_f0d96a2a27f0.slice - libcontainer container kubepods-besteffort-pod7d0ff7f0_5822_4b30_acc7_f0d96a2a27f0.slice. May 15 14:57:05.734040 systemd[1]: Created slice kubepods-besteffort-pod81d296b0_35d4_41fa_b814_4bd5b4888773.slice - libcontainer container kubepods-besteffort-pod81d296b0_35d4_41fa_b814_4bd5b4888773.slice. May 15 14:57:05.757268 systemd[1]: Created slice kubepods-besteffort-podf38d7bff_89c7_43fe_9821_67a2af7ff34c.slice - libcontainer container kubepods-besteffort-podf38d7bff_89c7_43fe_9821_67a2af7ff34c.slice. May 15 14:57:05.771038 systemd[1]: Created slice kubepods-besteffort-pod3300dd9b_523b_416e_9de8_89b5d9c22fa3.slice - libcontainer container kubepods-besteffort-pod3300dd9b_523b_416e_9de8_89b5d9c22fa3.slice. May 15 14:57:05.778354 containerd[1559]: time="2025-05-15T14:57:05.777700836Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5pnkc,Uid:3300dd9b-523b-416e-9de8-89b5d9c22fa3,Namespace:calico-system,Attempt:0,}" May 15 14:57:05.783950 kubelet[2683]: I0515 14:57:05.782080 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgr9\" (UniqueName: \"kubernetes.io/projected/e458056c-828c-4a75-acfb-1672e71f964a-kube-api-access-qwgr9\") pod \"coredns-668d6bf9bc-8gclb\" (UID: \"e458056c-828c-4a75-acfb-1672e71f964a\") " pod="kube-system/coredns-668d6bf9bc-8gclb" May 15 14:57:05.783950 kubelet[2683]: I0515 14:57:05.782144 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0-tigera-ca-bundle\") pod \"calico-kube-controllers-87fd6d8fb-2ntcg\" (UID: \"7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0\") " pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" May 15 14:57:05.783950 kubelet[2683]: I0515 14:57:05.782184 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/81d296b0-35d4-41fa-b814-4bd5b4888773-calico-apiserver-certs\") pod \"calico-apiserver-6684646c96-q9qjj\" (UID: \"81d296b0-35d4-41fa-b814-4bd5b4888773\") " pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" May 15 14:57:05.783950 kubelet[2683]: I0515 14:57:05.782211 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdwhx\" (UniqueName: \"kubernetes.io/projected/81d296b0-35d4-41fa-b814-4bd5b4888773-kube-api-access-wdwhx\") pod \"calico-apiserver-6684646c96-q9qjj\" (UID: \"81d296b0-35d4-41fa-b814-4bd5b4888773\") " pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" May 15 14:57:05.783950 kubelet[2683]: I0515 14:57:05.782247 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6ctf\" (UniqueName: \"kubernetes.io/projected/f38d7bff-89c7-43fe-9821-67a2af7ff34c-kube-api-access-j6ctf\") pod \"calico-apiserver-6684646c96-2mm5m\" (UID: \"f38d7bff-89c7-43fe-9821-67a2af7ff34c\") " pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" May 15 14:57:05.784230 kubelet[2683]: I0515 14:57:05.782285 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e458056c-828c-4a75-acfb-1672e71f964a-config-volume\") pod \"coredns-668d6bf9bc-8gclb\" (UID: \"e458056c-828c-4a75-acfb-1672e71f964a\") " pod="kube-system/coredns-668d6bf9bc-8gclb" May 15 14:57:05.784230 kubelet[2683]: I0515 14:57:05.782352 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thhq7\" (UniqueName: \"kubernetes.io/projected/7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0-kube-api-access-thhq7\") pod \"calico-kube-controllers-87fd6d8fb-2ntcg\" (UID: \"7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0\") " pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" May 15 14:57:05.784230 kubelet[2683]: I0515 14:57:05.782378 2683 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f38d7bff-89c7-43fe-9821-67a2af7ff34c-calico-apiserver-certs\") pod \"calico-apiserver-6684646c96-2mm5m\" (UID: \"f38d7bff-89c7-43fe-9821-67a2af7ff34c\") " pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" May 15 14:57:05.976211 kubelet[2683]: E0515 14:57:05.976166 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:05.979726 containerd[1559]: time="2025-05-15T14:57:05.979672746Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 15 14:57:05.992511 kubelet[2683]: E0515 14:57:05.992357 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:05.994948 containerd[1559]: time="2025-05-15T14:57:05.994892283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wbvvg,Uid:ed3eb677-7d86-464b-9053-ccf595c17dc1,Namespace:kube-system,Attempt:0,}" May 15 14:57:06.013420 kubelet[2683]: E0515 14:57:06.012961 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:06.016661 containerd[1559]: time="2025-05-15T14:57:06.016604244Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-8gclb,Uid:e458056c-828c-4a75-acfb-1672e71f964a,Namespace:kube-system,Attempt:0,}" May 15 14:57:06.024562 containerd[1559]: time="2025-05-15T14:57:06.024452398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-87fd6d8fb-2ntcg,Uid:7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0,Namespace:calico-system,Attempt:0,}" May 15 14:57:06.062700 containerd[1559]: time="2025-05-15T14:57:06.057829281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-q9qjj,Uid:81d296b0-35d4-41fa-b814-4bd5b4888773,Namespace:calico-apiserver,Attempt:0,}" May 15 14:57:06.077104 containerd[1559]: time="2025-05-15T14:57:06.076667838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-2mm5m,Uid:f38d7bff-89c7-43fe-9821-67a2af7ff34c,Namespace:calico-apiserver,Attempt:0,}" May 15 14:57:06.430673 containerd[1559]: time="2025-05-15T14:57:06.430617212Z" level=error msg="Failed to destroy network for sandbox \"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.442355 containerd[1559]: time="2025-05-15T14:57:06.442191432Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-8gclb,Uid:e458056c-828c-4a75-acfb-1672e71f964a,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.442760 containerd[1559]: time="2025-05-15T14:57:06.442628287Z" level=error msg="Failed to destroy network for sandbox \"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.442821 kubelet[2683]: E0515 14:57:06.442496 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.443474 kubelet[2683]: E0515 14:57:06.443341 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-8gclb" May 15 14:57:06.443474 kubelet[2683]: E0515 14:57:06.443449 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-8gclb" May 15 14:57:06.445606 kubelet[2683]: E0515 14:57:06.443948 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-8gclb_kube-system(e458056c-828c-4a75-acfb-1672e71f964a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-8gclb_kube-system(e458056c-828c-4a75-acfb-1672e71f964a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ef35a8a34da66170f470242d43731f8c5492256bfdd1802552c2b3a6f9643431\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-8gclb" podUID="e458056c-828c-4a75-acfb-1672e71f964a" May 15 14:57:06.445906 containerd[1559]: time="2025-05-15T14:57:06.445851012Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-2mm5m,Uid:f38d7bff-89c7-43fe-9821-67a2af7ff34c,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.448122 kubelet[2683]: E0515 14:57:06.447497 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.448122 kubelet[2683]: E0515 14:57:06.447569 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" May 15 14:57:06.448122 kubelet[2683]: E0515 14:57:06.447622 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" May 15 14:57:06.448446 kubelet[2683]: E0515 14:57:06.447674 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6684646c96-2mm5m_calico-apiserver(f38d7bff-89c7-43fe-9821-67a2af7ff34c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6684646c96-2mm5m_calico-apiserver(f38d7bff-89c7-43fe-9821-67a2af7ff34c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"580b0b8b594ee0eac22e91622b97460626b0037611f805ee93ef61844f411ed9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" podUID="f38d7bff-89c7-43fe-9821-67a2af7ff34c" May 15 14:57:06.454817 containerd[1559]: time="2025-05-15T14:57:06.454729016Z" level=error msg="Failed to destroy network for sandbox \"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.460861 containerd[1559]: time="2025-05-15T14:57:06.460747625Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5pnkc,Uid:3300dd9b-523b-416e-9de8-89b5d9c22fa3,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.462138 kubelet[2683]: E0515 14:57:06.461044 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.462138 kubelet[2683]: E0515 14:57:06.461119 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5pnkc" May 15 14:57:06.462138 kubelet[2683]: E0515 14:57:06.461150 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5pnkc" May 15 14:57:06.462609 kubelet[2683]: E0515 14:57:06.461207 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5pnkc_calico-system(3300dd9b-523b-416e-9de8-89b5d9c22fa3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5pnkc_calico-system(3300dd9b-523b-416e-9de8-89b5d9c22fa3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8844803a14a965f084d928b2ba5c9c5c7f8039f863c3bfa34607513182102a34\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5pnkc" podUID="3300dd9b-523b-416e-9de8-89b5d9c22fa3" May 15 14:57:06.482119 containerd[1559]: time="2025-05-15T14:57:06.482065095Z" level=error msg="Failed to destroy network for sandbox \"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.483321 containerd[1559]: time="2025-05-15T14:57:06.483244695Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-q9qjj,Uid:81d296b0-35d4-41fa-b814-4bd5b4888773,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.483924 kubelet[2683]: E0515 14:57:06.483535 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.483924 kubelet[2683]: E0515 14:57:06.483617 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" May 15 14:57:06.483924 kubelet[2683]: E0515 14:57:06.483642 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" May 15 14:57:06.484093 kubelet[2683]: E0515 14:57:06.483695 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6684646c96-q9qjj_calico-apiserver(81d296b0-35d4-41fa-b814-4bd5b4888773)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6684646c96-q9qjj_calico-apiserver(81d296b0-35d4-41fa-b814-4bd5b4888773)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5fa17daafd4451c9f32f4684e7ce3a8ca12b0189fe826571022e0571c61d4dba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" podUID="81d296b0-35d4-41fa-b814-4bd5b4888773" May 15 14:57:06.495301 containerd[1559]: time="2025-05-15T14:57:06.495237242Z" level=error msg="Failed to destroy network for sandbox \"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.497045 containerd[1559]: time="2025-05-15T14:57:06.496845950Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-87fd6d8fb-2ntcg,Uid:7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.497437 kubelet[2683]: E0515 14:57:06.497384 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.497814 kubelet[2683]: E0515 14:57:06.497468 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" May 15 14:57:06.497814 kubelet[2683]: E0515 14:57:06.497496 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" May 15 14:57:06.497814 kubelet[2683]: E0515 14:57:06.497609 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-87fd6d8fb-2ntcg_calico-system(7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-87fd6d8fb-2ntcg_calico-system(7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5630b4aaab312232c1b4e3ce020e46697f839822becbb068fb1ef71a2bbec2d8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" podUID="7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0" May 15 14:57:06.520744 containerd[1559]: time="2025-05-15T14:57:06.520643532Z" level=error msg="Failed to destroy network for sandbox \"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.521670 containerd[1559]: time="2025-05-15T14:57:06.521612558Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wbvvg,Uid:ed3eb677-7d86-464b-9053-ccf595c17dc1,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.522108 kubelet[2683]: E0515 14:57:06.522056 2683 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 15 14:57:06.522297 kubelet[2683]: E0515 14:57:06.522230 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-wbvvg" May 15 14:57:06.522376 kubelet[2683]: E0515 14:57:06.522258 2683 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-wbvvg" May 15 14:57:06.522508 kubelet[2683]: E0515 14:57:06.522476 2683 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-wbvvg_kube-system(ed3eb677-7d86-464b-9053-ccf595c17dc1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-wbvvg_kube-system(ed3eb677-7d86-464b-9053-ccf595c17dc1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d238576400e7dca9720f06f71a918fed6a98ce850b8dc927433a2e2b8e38bb2b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-wbvvg" podUID="ed3eb677-7d86-464b-9053-ccf595c17dc1" May 15 14:57:06.799502 systemd[1]: run-netns-cni\x2dec99ef48\x2d6d94\x2d1f93\x2de522\x2d1ce23fc8a012.mount: Deactivated successfully. May 15 14:57:14.811639 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2947882665.mount: Deactivated successfully. May 15 14:57:14.853614 containerd[1559]: time="2025-05-15T14:57:14.853126718Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:14.873008 containerd[1559]: time="2025-05-15T14:57:14.858397961Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=144068748" May 15 14:57:14.873309 containerd[1559]: time="2025-05-15T14:57:14.859189127Z" level=info msg="ImageCreate event name:\"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:14.873442 containerd[1559]: time="2025-05-15T14:57:14.862214159Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"144068610\" in 8.882492961s" May 15 14:57:14.873544 containerd[1559]: time="2025-05-15T14:57:14.873517740Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" May 15 14:57:14.876130 containerd[1559]: time="2025-05-15T14:57:14.876043704Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:14.921163 containerd[1559]: time="2025-05-15T14:57:14.921091934Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 15 14:57:14.933256 containerd[1559]: time="2025-05-15T14:57:14.933195374Z" level=info msg="Container 8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:14.948219 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1714720978.mount: Deactivated successfully. May 15 14:57:14.993624 containerd[1559]: time="2025-05-15T14:57:14.993517745Z" level=info msg="CreateContainer within sandbox \"4537d9267232f81946b23e0bb906d31a8ef792093beabc0152c911b30f4326b1\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\"" May 15 14:57:14.996125 containerd[1559]: time="2025-05-15T14:57:14.996031345Z" level=info msg="StartContainer for \"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\"" May 15 14:57:15.001981 containerd[1559]: time="2025-05-15T14:57:15.001693970Z" level=info msg="connecting to shim 8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f" address="unix:///run/containerd/s/171f23b7b14ba40beb4a1e82dcc0e417284b841849914f11bd47021ec53e26f0" protocol=ttrpc version=3 May 15 14:57:15.129933 systemd[1]: Started cri-containerd-8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f.scope - libcontainer container 8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f. May 15 14:57:15.237354 containerd[1559]: time="2025-05-15T14:57:15.237107063Z" level=info msg="StartContainer for \"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" returns successfully" May 15 14:57:15.531150 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 15 14:57:15.531313 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 15 14:57:16.068866 kubelet[2683]: E0515 14:57:16.068784 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:16.311168 containerd[1559]: time="2025-05-15T14:57:16.310973001Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"ff9fe7876cc34b9036cee121f65caff1e5b8f78abb509bceb8549e638c6b695d\" pid:3652 exit_status:1 exited_at:{seconds:1747321036 nanos:310412801}" May 15 14:57:17.071885 kubelet[2683]: E0515 14:57:17.071774 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:17.243554 containerd[1559]: time="2025-05-15T14:57:17.243387652Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"efa22c124839e902caaa4dce2ee91ab145d4d5e0c16a1f2ec1380461d092967c\" pid:3676 exit_status:1 exited_at:{seconds:1747321037 nanos:242439897}" May 15 14:57:17.732128 containerd[1559]: time="2025-05-15T14:57:17.731881783Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-2mm5m,Uid:f38d7bff-89c7-43fe-9821-67a2af7ff34c,Namespace:calico-apiserver,Attempt:0,}" May 15 14:57:18.093048 kubelet[2683]: E0515 14:57:18.092967 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:18.255871 systemd-networkd[1431]: calie243c8f7f89: Link UP May 15 14:57:18.258284 systemd-networkd[1431]: calie243c8f7f89: Gained carrier May 15 14:57:18.261826 containerd[1559]: time="2025-05-15T14:57:18.261174379Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"ac29f149695b002849e4a3521208d87b91edfab381ef6ce37d9bac012832e4c4\" pid:3815 exit_status:1 exited_at:{seconds:1747321038 nanos:259139445}" May 15 14:57:18.291554 kubelet[2683]: I0515 14:57:18.290581 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-7c27f" podStartSLOduration=4.65727623 podStartE2EDuration="25.289370609s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:56:54.252507342 +0000 UTC m=+11.755200236" lastFinishedPulling="2025-05-15 14:57:14.884601721 +0000 UTC m=+32.387294615" observedRunningTime="2025-05-15 14:57:16.111497485 +0000 UTC m=+33.614190382" watchObservedRunningTime="2025-05-15 14:57:18.289370609 +0000 UTC m=+35.792063511" May 15 14:57:18.298929 containerd[1559]: 2025-05-15 14:57:17.816 [INFO][3784] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 14:57:18.298929 containerd[1559]: 2025-05-15 14:57:17.866 [INFO][3784] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0 calico-apiserver-6684646c96- calico-apiserver f38d7bff-89c7-43fe-9821-67a2af7ff34c 691 0 2025-05-15 14:56:53 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6684646c96 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f calico-apiserver-6684646c96-2mm5m eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calie243c8f7f89 [] []}} ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-" May 15 14:57:18.298929 containerd[1559]: 2025-05-15 14:57:17.867 [INFO][3784] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.298929 containerd[1559]: 2025-05-15 14:57:18.139 [INFO][3795] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" HandleID="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.169 [INFO][3795] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" HandleID="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d3700), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"calico-apiserver-6684646c96-2mm5m", "timestamp":"2025-05-15 14:57:18.139677567 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.170 [INFO][3795] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.170 [INFO][3795] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.170 [INFO][3795] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.177 [INFO][3795] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.191 [INFO][3795] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.202 [INFO][3795] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.206 [INFO][3795] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299701 containerd[1559]: 2025-05-15 14:57:18.212 [INFO][3795] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.212 [INFO][3795] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.215 [INFO][3795] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659 May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.222 [INFO][3795] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.230 [INFO][3795] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.193/26] block=192.168.95.192/26 handle="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.231 [INFO][3795] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.193/26] handle="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.231 [INFO][3795] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:18.299960 containerd[1559]: 2025-05-15 14:57:18.231 [INFO][3795] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.193/26] IPv6=[] ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" HandleID="k8s-pod-network.bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.300141 containerd[1559]: 2025-05-15 14:57:18.236 [INFO][3784] cni-plugin/k8s.go 386: Populated endpoint ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0", GenerateName:"calico-apiserver-6684646c96-", Namespace:"calico-apiserver", SelfLink:"", UID:"f38d7bff-89c7-43fe-9821-67a2af7ff34c", ResourceVersion:"691", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6684646c96", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"calico-apiserver-6684646c96-2mm5m", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.95.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie243c8f7f89", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:18.300236 containerd[1559]: 2025-05-15 14:57:18.236 [INFO][3784] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.193/32] ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.300236 containerd[1559]: 2025-05-15 14:57:18.238 [INFO][3784] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie243c8f7f89 ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.300236 containerd[1559]: 2025-05-15 14:57:18.253 [INFO][3784] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.300330 containerd[1559]: 2025-05-15 14:57:18.254 [INFO][3784] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0", GenerateName:"calico-apiserver-6684646c96-", Namespace:"calico-apiserver", SelfLink:"", UID:"f38d7bff-89c7-43fe-9821-67a2af7ff34c", ResourceVersion:"691", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6684646c96", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659", Pod:"calico-apiserver-6684646c96-2mm5m", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.95.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie243c8f7f89", MAC:"12:80:68:47:20:8a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:18.300411 containerd[1559]: 2025-05-15 14:57:18.281 [INFO][3784] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-2mm5m" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--2mm5m-eth0" May 15 14:57:18.475696 containerd[1559]: time="2025-05-15T14:57:18.471214976Z" level=info msg="connecting to shim bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659" address="unix:///run/containerd/s/1beb675a707e59bbe9e3eb7e0a260e77384a546b544e17ddde5ae940f2d64bb9" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:18.524786 systemd[1]: Started cri-containerd-bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659.scope - libcontainer container bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659. May 15 14:57:18.656403 containerd[1559]: time="2025-05-15T14:57:18.656220431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-2mm5m,Uid:f38d7bff-89c7-43fe-9821-67a2af7ff34c,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659\"" May 15 14:57:18.662622 containerd[1559]: time="2025-05-15T14:57:18.662569858Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 15 14:57:19.185564 kubelet[2683]: I0515 14:57:19.185355 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 14:57:19.187548 kubelet[2683]: E0515 14:57:19.187461 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:19.475782 systemd-networkd[1431]: calie243c8f7f89: Gained IPv6LL May 15 14:57:19.709829 containerd[1559]: time="2025-05-15T14:57:19.709755975Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-87fd6d8fb-2ntcg,Uid:7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0,Namespace:calico-system,Attempt:0,}" May 15 14:57:19.711602 containerd[1559]: time="2025-05-15T14:57:19.710810513Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-8gclb,Uid:e458056c-828c-4a75-acfb-1672e71f964a,Namespace:kube-system,Attempt:0,}" May 15 14:57:19.712048 kubelet[2683]: E0515 14:57:19.709877 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:20.100503 kubelet[2683]: E0515 14:57:20.100461 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:20.128216 systemd-networkd[1431]: cali423e79a300c: Link UP May 15 14:57:20.132145 systemd-networkd[1431]: cali423e79a300c: Gained carrier May 15 14:57:20.170134 containerd[1559]: 2025-05-15 14:57:19.878 [INFO][3920] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 14:57:20.170134 containerd[1559]: 2025-05-15 14:57:19.909 [INFO][3920] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0 coredns-668d6bf9bc- kube-system e458056c-828c-4a75-acfb-1672e71f964a 689 0 2025-05-15 14:56:47 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f coredns-668d6bf9bc-8gclb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali423e79a300c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-" May 15 14:57:20.170134 containerd[1559]: 2025-05-15 14:57:19.910 [INFO][3920] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.170134 containerd[1559]: 2025-05-15 14:57:19.999 [INFO][3952] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" HandleID="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.030 [INFO][3952] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" HandleID="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000500f0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"coredns-668d6bf9bc-8gclb", "timestamp":"2025-05-15 14:57:19.998599413 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.031 [INFO][3952] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.031 [INFO][3952] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.031 [INFO][3952] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.035 [INFO][3952] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.049 [INFO][3952] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.061 [INFO][3952] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.066 [INFO][3952] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170604 containerd[1559]: 2025-05-15 14:57:20.072 [INFO][3952] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.072 [INFO][3952] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.077 [INFO][3952] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082 May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.089 [INFO][3952] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.103 [INFO][3952] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.194/26] block=192.168.95.192/26 handle="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.103 [INFO][3952] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.194/26] handle="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.104 [INFO][3952] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:20.170996 containerd[1559]: 2025-05-15 14:57:20.104 [INFO][3952] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.194/26] IPv6=[] ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" HandleID="k8s-pod-network.f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.115 [INFO][3920] cni-plugin/k8s.go 386: Populated endpoint ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e458056c-828c-4a75-acfb-1672e71f964a", ResourceVersion:"689", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"coredns-668d6bf9bc-8gclb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.95.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali423e79a300c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.116 [INFO][3920] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.194/32] ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.117 [INFO][3920] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali423e79a300c ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.136 [INFO][3920] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.138 [INFO][3920] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"e458056c-828c-4a75-acfb-1672e71f964a", ResourceVersion:"689", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082", Pod:"coredns-668d6bf9bc-8gclb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.95.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali423e79a300c", MAC:"f2:71:d5:ed:35:a1", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:20.172566 containerd[1559]: 2025-05-15 14:57:20.161 [INFO][3920] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" Namespace="kube-system" Pod="coredns-668d6bf9bc-8gclb" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--8gclb-eth0" May 15 14:57:20.228222 containerd[1559]: time="2025-05-15T14:57:20.228161419Z" level=info msg="connecting to shim f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082" address="unix:///run/containerd/s/2da1a56fda65c70f6e41f1f61ce7a4890e17191f2dc9ba8a959810067d762cf4" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:20.314619 systemd-networkd[1431]: calibfd0c29151e: Link UP May 15 14:57:20.315630 systemd-networkd[1431]: calibfd0c29151e: Gained carrier May 15 14:57:20.335218 systemd[1]: Started cri-containerd-f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082.scope - libcontainer container f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082. May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:19.878 [INFO][3917] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:19.917 [INFO][3917] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0 calico-kube-controllers-87fd6d8fb- calico-system 7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0 688 0 2025-05-15 14:56:53 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:87fd6d8fb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f calico-kube-controllers-87fd6d8fb-2ntcg eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calibfd0c29151e [] []}} ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:19.917 [INFO][3917] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.074 [INFO][3957] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" HandleID="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.128 [INFO][3957] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" HandleID="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00040dae0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"calico-kube-controllers-87fd6d8fb-2ntcg", "timestamp":"2025-05-15 14:57:20.07444754 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.128 [INFO][3957] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.128 [INFO][3957] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.128 [INFO][3957] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.143 [INFO][3957] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.223 [INFO][3957] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.238 [INFO][3957] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.243 [INFO][3957] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.250 [INFO][3957] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.250 [INFO][3957] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.258 [INFO][3957] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.279 [INFO][3957] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.294 [INFO][3957] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.195/26] block=192.168.95.192/26 handle="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.294 [INFO][3957] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.195/26] handle="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.294 [INFO][3957] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:20.368434 containerd[1559]: 2025-05-15 14:57:20.294 [INFO][3957] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.195/26] IPv6=[] ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" HandleID="k8s-pod-network.935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.306 [INFO][3917] cni-plugin/k8s.go 386: Populated endpoint ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0", GenerateName:"calico-kube-controllers-87fd6d8fb-", Namespace:"calico-system", SelfLink:"", UID:"7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0", ResourceVersion:"688", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"87fd6d8fb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"calico-kube-controllers-87fd6d8fb-2ntcg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.95.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calibfd0c29151e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.307 [INFO][3917] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.195/32] ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.307 [INFO][3917] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibfd0c29151e ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.316 [INFO][3917] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.317 [INFO][3917] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0", GenerateName:"calico-kube-controllers-87fd6d8fb-", Namespace:"calico-system", SelfLink:"", UID:"7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0", ResourceVersion:"688", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"87fd6d8fb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf", Pod:"calico-kube-controllers-87fd6d8fb-2ntcg", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.95.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calibfd0c29151e", MAC:"8a:97:83:2a:8b:a2", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:20.369107 containerd[1559]: 2025-05-15 14:57:20.362 [INFO][3917] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" Namespace="calico-system" Pod="calico-kube-controllers-87fd6d8fb-2ntcg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--kube--controllers--87fd6d8fb--2ntcg-eth0" May 15 14:57:20.442778 containerd[1559]: time="2025-05-15T14:57:20.442704444Z" level=info msg="connecting to shim 935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf" address="unix:///run/containerd/s/3ff5a9413315124770f10298d98cc053e12a415b2bcf48e53eef81ca70d00104" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:20.568193 containerd[1559]: time="2025-05-15T14:57:20.567168786Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-8gclb,Uid:e458056c-828c-4a75-acfb-1672e71f964a,Namespace:kube-system,Attempt:0,} returns sandbox id \"f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082\"" May 15 14:57:20.571046 kubelet[2683]: E0515 14:57:20.570720 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:20.579295 containerd[1559]: time="2025-05-15T14:57:20.577430807Z" level=info msg="CreateContainer within sandbox \"f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 15 14:57:20.605957 systemd[1]: Started cri-containerd-935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf.scope - libcontainer container 935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf. May 15 14:57:20.632290 containerd[1559]: time="2025-05-15T14:57:20.630259971Z" level=info msg="Container ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:20.641694 containerd[1559]: time="2025-05-15T14:57:20.641644914Z" level=info msg="CreateContainer within sandbox \"f5223d3f5691037256e3a6b0653615c374b626c61a71920db18a586f01e69082\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4\"" May 15 14:57:20.644213 containerd[1559]: time="2025-05-15T14:57:20.644149886Z" level=info msg="StartContainer for \"ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4\"" May 15 14:57:20.645504 containerd[1559]: time="2025-05-15T14:57:20.645370828Z" level=info msg="connecting to shim ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4" address="unix:///run/containerd/s/2da1a56fda65c70f6e41f1f61ce7a4890e17191f2dc9ba8a959810067d762cf4" protocol=ttrpc version=3 May 15 14:57:20.695902 systemd[1]: Started cri-containerd-ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4.scope - libcontainer container ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4. May 15 14:57:20.711341 containerd[1559]: time="2025-05-15T14:57:20.710867724Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-q9qjj,Uid:81d296b0-35d4-41fa-b814-4bd5b4888773,Namespace:calico-apiserver,Attempt:0,}" May 15 14:57:20.842861 containerd[1559]: time="2025-05-15T14:57:20.842518852Z" level=info msg="StartContainer for \"ff02855cb68d7fc47937541f3cee438ba1e65c9fdddd65ca31124c89dfbfd2f4\" returns successfully" May 15 14:57:21.081673 containerd[1559]: time="2025-05-15T14:57:21.081180891Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-87fd6d8fb-2ntcg,Uid:7d0ff7f0-5822-4b30-acc7-f0d96a2a27f0,Namespace:calico-system,Attempt:0,} returns sandbox id \"935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf\"" May 15 14:57:21.205964 kubelet[2683]: E0515 14:57:21.205800 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:21.333871 systemd-networkd[1431]: cali48c79441d70: Link UP May 15 14:57:21.339054 systemd-networkd[1431]: cali48c79441d70: Gained carrier May 15 14:57:21.373709 kubelet[2683]: I0515 14:57:21.373377 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-8gclb" podStartSLOduration=34.373350882 podStartE2EDuration="34.373350882s" podCreationTimestamp="2025-05-15 14:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:57:21.25698898 +0000 UTC m=+38.759681880" watchObservedRunningTime="2025-05-15 14:57:21.373350882 +0000 UTC m=+38.876043775" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:20.934 [INFO][4128] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0 calico-apiserver-6684646c96- calico-apiserver 81d296b0-35d4-41fa-b814-4bd5b4888773 690 0 2025-05-15 14:56:53 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6684646c96 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f calico-apiserver-6684646c96-q9qjj eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali48c79441d70 [] []}} ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:20.934 [INFO][4128] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.077 [INFO][4155] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" HandleID="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.121 [INFO][4155] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" HandleID="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003a55d0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"calico-apiserver-6684646c96-q9qjj", "timestamp":"2025-05-15 14:57:21.076997187 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.121 [INFO][4155] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.121 [INFO][4155] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.121 [INFO][4155] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.131 [INFO][4155] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.153 [INFO][4155] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.181 [INFO][4155] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.191 [INFO][4155] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.214 [INFO][4155] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.215 [INFO][4155] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.225 [INFO][4155] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.277 [INFO][4155] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.304 [INFO][4155] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.196/26] block=192.168.95.192/26 handle="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.305 [INFO][4155] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.196/26] handle="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.305 [INFO][4155] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:21.383166 containerd[1559]: 2025-05-15 14:57:21.305 [INFO][4155] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.196/26] IPv6=[] ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" HandleID="k8s-pod-network.1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Workload="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.318 [INFO][4128] cni-plugin/k8s.go 386: Populated endpoint ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0", GenerateName:"calico-apiserver-6684646c96-", Namespace:"calico-apiserver", SelfLink:"", UID:"81d296b0-35d4-41fa-b814-4bd5b4888773", ResourceVersion:"690", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6684646c96", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"calico-apiserver-6684646c96-q9qjj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.95.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali48c79441d70", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.318 [INFO][4128] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.196/32] ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.319 [INFO][4128] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali48c79441d70 ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.339 [INFO][4128] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.343 [INFO][4128] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0", GenerateName:"calico-apiserver-6684646c96-", Namespace:"calico-apiserver", SelfLink:"", UID:"81d296b0-35d4-41fa-b814-4bd5b4888773", ResourceVersion:"690", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6684646c96", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d", Pod:"calico-apiserver-6684646c96-q9qjj", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.95.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali48c79441d70", MAC:"76:7d:30:27:e2:e0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:21.385260 containerd[1559]: 2025-05-15 14:57:21.371 [INFO][4128] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" Namespace="calico-apiserver" Pod="calico-apiserver-6684646c96-q9qjj" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-calico--apiserver--6684646c96--q9qjj-eth0" May 15 14:57:21.397331 systemd-networkd[1431]: calibfd0c29151e: Gained IPv6LL May 15 14:57:21.462229 containerd[1559]: time="2025-05-15T14:57:21.462101367Z" level=info msg="connecting to shim 1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d" address="unix:///run/containerd/s/f4f79168047c2f927091ff7a0a4911abdb77107447b14acbf28aafc3583dc703" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:21.566892 systemd[1]: Started cri-containerd-1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d.scope - libcontainer container 1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d. May 15 14:57:21.602288 systemd-networkd[1431]: vxlan.calico: Link UP May 15 14:57:21.604037 systemd-networkd[1431]: vxlan.calico: Gained carrier May 15 14:57:21.713326 kubelet[2683]: E0515 14:57:21.713294 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:21.717506 containerd[1559]: time="2025-05-15T14:57:21.716923192Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5pnkc,Uid:3300dd9b-523b-416e-9de8-89b5d9c22fa3,Namespace:calico-system,Attempt:0,}" May 15 14:57:21.720998 containerd[1559]: time="2025-05-15T14:57:21.720848091Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wbvvg,Uid:ed3eb677-7d86-464b-9053-ccf595c17dc1,Namespace:kube-system,Attempt:0,}" May 15 14:57:21.869313 containerd[1559]: time="2025-05-15T14:57:21.869061521Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6684646c96-q9qjj,Uid:81d296b0-35d4-41fa-b814-4bd5b4888773,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d\"" May 15 14:57:22.101185 systemd-networkd[1431]: cali423e79a300c: Gained IPv6LL May 15 14:57:22.240629 kubelet[2683]: E0515 14:57:22.239167 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:22.261162 systemd-networkd[1431]: calidcfccc5267c: Link UP May 15 14:57:22.269604 systemd-networkd[1431]: calidcfccc5267c: Gained carrier May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:21.994 [INFO][4269] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0 coredns-668d6bf9bc- kube-system ed3eb677-7d86-464b-9053-ccf595c17dc1 684 0 2025-05-15 14:56:47 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f coredns-668d6bf9bc-wbvvg eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calidcfccc5267c [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:21.995 [INFO][4269] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.104 [INFO][4304] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" HandleID="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.126 [INFO][4304] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" HandleID="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002b3f50), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"coredns-668d6bf9bc-wbvvg", "timestamp":"2025-05-15 14:57:22.104196216 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.129 [INFO][4304] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.129 [INFO][4304] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.130 [INFO][4304] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.144 [INFO][4304] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.163 [INFO][4304] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.175 [INFO][4304] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.180 [INFO][4304] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.188 [INFO][4304] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.188 [INFO][4304] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.196 [INFO][4304] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228 May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.209 [INFO][4304] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.232 [INFO][4304] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.197/26] block=192.168.95.192/26 handle="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.232 [INFO][4304] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.197/26] handle="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.232 [INFO][4304] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:22.340414 containerd[1559]: 2025-05-15 14:57:22.233 [INFO][4304] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.197/26] IPv6=[] ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" HandleID="k8s-pod-network.dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Workload="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.245 [INFO][4269] cni-plugin/k8s.go 386: Populated endpoint ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ed3eb677-7d86-464b-9053-ccf595c17dc1", ResourceVersion:"684", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"coredns-668d6bf9bc-wbvvg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.95.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calidcfccc5267c", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.247 [INFO][4269] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.197/32] ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.247 [INFO][4269] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calidcfccc5267c ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.273 [INFO][4269] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.276 [INFO][4269] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"ed3eb677-7d86-464b-9053-ccf595c17dc1", ResourceVersion:"684", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228", Pod:"coredns-668d6bf9bc-wbvvg", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.95.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calidcfccc5267c", MAC:"d2:6c:71:1b:d9:ce", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:22.343440 containerd[1559]: 2025-05-15 14:57:22.329 [INFO][4269] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" Namespace="kube-system" Pod="coredns-668d6bf9bc-wbvvg" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-coredns--668d6bf9bc--wbvvg-eth0" May 15 14:57:22.420607 containerd[1559]: time="2025-05-15T14:57:22.420051666Z" level=info msg="connecting to shim dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228" address="unix:///run/containerd/s/13a4c076fdfa61d55c9c959a4eab19a557c1bf4d1f8c3dfe94ea18580e681ed2" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:22.483115 systemd-networkd[1431]: cali6e5c00a4df4: Link UP May 15 14:57:22.484069 systemd-networkd[1431]: cali6e5c00a4df4: Gained carrier May 15 14:57:22.501937 systemd[1]: Started cri-containerd-dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228.scope - libcontainer container dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228. May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:21.951 [INFO][4276] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0 csi-node-driver- calico-system 3300dd9b-523b-416e-9de8-89b5d9c22fa3 595 0 2025-05-15 14:56:53 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5b5cc68cd5 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4334.0.0-a-c1acadba3f csi-node-driver-5pnkc eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali6e5c00a4df4 [] []}} ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:21.951 [INFO][4276] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.151 [INFO][4299] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" HandleID="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Workload="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.177 [INFO][4299] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" HandleID="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Workload="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000590ec0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4334.0.0-a-c1acadba3f", "pod":"csi-node-driver-5pnkc", "timestamp":"2025-05-15 14:57:22.148482415 +0000 UTC"}, Hostname:"ci-4334.0.0-a-c1acadba3f", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.178 [INFO][4299] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.232 [INFO][4299] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.232 [INFO][4299] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4334.0.0-a-c1acadba3f' May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.245 [INFO][4299] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.348 [INFO][4299] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.379 [INFO][4299] ipam/ipam.go 489: Trying affinity for 192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.388 [INFO][4299] ipam/ipam.go 155: Attempting to load block cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.412 [INFO][4299] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.95.192/26 host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.414 [INFO][4299] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.95.192/26 handle="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.427 [INFO][4299] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6 May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.448 [INFO][4299] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.95.192/26 handle="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.468 [INFO][4299] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.95.198/26] block=192.168.95.192/26 handle="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.469 [INFO][4299] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.95.198/26] handle="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" host="ci-4334.0.0-a-c1acadba3f" May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.469 [INFO][4299] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 15 14:57:22.541825 containerd[1559]: 2025-05-15 14:57:22.469 [INFO][4299] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.95.198/26] IPv6=[] ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" HandleID="k8s-pod-network.686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Workload="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.476 [INFO][4276] cni-plugin/k8s.go 386: Populated endpoint ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"3300dd9b-523b-416e-9de8-89b5d9c22fa3", ResourceVersion:"595", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"", Pod:"csi-node-driver-5pnkc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.95.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6e5c00a4df4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.476 [INFO][4276] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.95.198/32] ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.476 [INFO][4276] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6e5c00a4df4 ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.484 [INFO][4276] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.485 [INFO][4276] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"3300dd9b-523b-416e-9de8-89b5d9c22fa3", ResourceVersion:"595", Generation:0, CreationTimestamp:time.Date(2025, time.May, 15, 14, 56, 53, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4334.0.0-a-c1acadba3f", ContainerID:"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6", Pod:"csi-node-driver-5pnkc", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.95.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6e5c00a4df4", MAC:"3a:d3:d5:6f:01:35", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 15 14:57:22.544253 containerd[1559]: 2025-05-15 14:57:22.524 [INFO][4276] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" Namespace="calico-system" Pod="csi-node-driver-5pnkc" WorkloadEndpoint="ci--4334.0.0--a--c1acadba3f-k8s-csi--node--driver--5pnkc-eth0" May 15 14:57:22.608841 containerd[1559]: time="2025-05-15T14:57:22.608794234Z" level=info msg="connecting to shim 686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6" address="unix:///run/containerd/s/178158c20a30b24b99dfdf5e8426e1a5423b7658001a95f16bc96aa6164049d2" namespace=k8s.io protocol=ttrpc version=3 May 15 14:57:22.663763 systemd[1]: Started cri-containerd-686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6.scope - libcontainer container 686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6. May 15 14:57:22.675882 systemd-networkd[1431]: cali48c79441d70: Gained IPv6LL May 15 14:57:22.699765 containerd[1559]: time="2025-05-15T14:57:22.699725605Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wbvvg,Uid:ed3eb677-7d86-464b-9053-ccf595c17dc1,Namespace:kube-system,Attempt:0,} returns sandbox id \"dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228\"" May 15 14:57:22.703703 kubelet[2683]: E0515 14:57:22.703674 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:22.723321 containerd[1559]: time="2025-05-15T14:57:22.721861715Z" level=info msg="CreateContainer within sandbox \"dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 15 14:57:22.803131 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3732126894.mount: Deactivated successfully. May 15 14:57:22.817890 containerd[1559]: time="2025-05-15T14:57:22.817836300Z" level=info msg="Container 5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:22.845999 containerd[1559]: time="2025-05-15T14:57:22.844464053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5pnkc,Uid:3300dd9b-523b-416e-9de8-89b5d9c22fa3,Namespace:calico-system,Attempt:0,} returns sandbox id \"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6\"" May 15 14:57:22.887753 containerd[1559]: time="2025-05-15T14:57:22.886924748Z" level=info msg="CreateContainer within sandbox \"dd1c57885e2eec97ed7b718e8c34a62fd2f31a4c2192b3b4584587167c29d228\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37\"" May 15 14:57:22.890931 containerd[1559]: time="2025-05-15T14:57:22.890862433Z" level=info msg="StartContainer for \"5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37\"" May 15 14:57:22.892276 containerd[1559]: time="2025-05-15T14:57:22.892206710Z" level=info msg="connecting to shim 5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37" address="unix:///run/containerd/s/13a4c076fdfa61d55c9c959a4eab19a557c1bf4d1f8c3dfe94ea18580e681ed2" protocol=ttrpc version=3 May 15 14:57:22.990967 systemd[1]: Started cri-containerd-5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37.scope - libcontainer container 5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37. May 15 14:57:23.144639 containerd[1559]: time="2025-05-15T14:57:23.144404930Z" level=info msg="StartContainer for \"5b719ac558409bab007ae38b1ded2714f1fb337818aea8b6011d9f9e0831fe37\" returns successfully" May 15 14:57:23.189620 systemd-networkd[1431]: vxlan.calico: Gained IPv6LL May 15 14:57:23.263391 kubelet[2683]: E0515 14:57:23.263342 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:23.267552 kubelet[2683]: E0515 14:57:23.266235 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:23.764679 systemd-networkd[1431]: calidcfccc5267c: Gained IPv6LL May 15 14:57:23.806867 containerd[1559]: time="2025-05-15T14:57:23.806795399Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:23.807758 containerd[1559]: time="2025-05-15T14:57:23.807711007Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=43021437" May 15 14:57:23.818852 containerd[1559]: time="2025-05-15T14:57:23.818788808Z" level=info msg="ImageCreate event name:\"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:23.823562 containerd[1559]: time="2025-05-15T14:57:23.822830490Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:23.826951 containerd[1559]: time="2025-05-15T14:57:23.826872198Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 5.159387145s" May 15 14:57:23.827239 containerd[1559]: time="2025-05-15T14:57:23.827208093Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 15 14:57:23.829674 containerd[1559]: time="2025-05-15T14:57:23.829630470Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 15 14:57:23.832587 containerd[1559]: time="2025-05-15T14:57:23.831480300Z" level=info msg="CreateContainer within sandbox \"bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 14:57:23.842575 containerd[1559]: time="2025-05-15T14:57:23.841659004Z" level=info msg="Container fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:23.870152 containerd[1559]: time="2025-05-15T14:57:23.869356792Z" level=info msg="CreateContainer within sandbox \"bff2e8321afa08032990349acda3240857cac8981145fc809207a8ac7fdd6659\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a\"" May 15 14:57:23.872875 containerd[1559]: time="2025-05-15T14:57:23.872826149Z" level=info msg="StartContainer for \"fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a\"" May 15 14:57:23.912621 containerd[1559]: time="2025-05-15T14:57:23.912561488Z" level=info msg="connecting to shim fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a" address="unix:///run/containerd/s/1beb675a707e59bbe9e3eb7e0a260e77384a546b544e17ddde5ae940f2d64bb9" protocol=ttrpc version=3 May 15 14:57:23.955936 systemd-networkd[1431]: cali6e5c00a4df4: Gained IPv6LL May 15 14:57:23.976248 systemd[1]: Started cri-containerd-fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a.scope - libcontainer container fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a. May 15 14:57:24.087501 containerd[1559]: time="2025-05-15T14:57:24.087301795Z" level=info msg="StartContainer for \"fc83abba5765da0c0beecf6badc1a278ab090715df957b78a25bc13b0d07835a\" returns successfully" May 15 14:57:24.270232 kubelet[2683]: E0515 14:57:24.269512 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:24.271102 kubelet[2683]: E0515 14:57:24.271042 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:24.312151 kubelet[2683]: I0515 14:57:24.312057 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-wbvvg" podStartSLOduration=37.312022601 podStartE2EDuration="37.312022601s" podCreationTimestamp="2025-05-15 14:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-15 14:57:23.298227358 +0000 UTC m=+40.800920252" watchObservedRunningTime="2025-05-15 14:57:24.312022601 +0000 UTC m=+41.814715509" May 15 14:57:24.312579 kubelet[2683]: I0515 14:57:24.312501 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6684646c96-2mm5m" podStartSLOduration=26.144111737 podStartE2EDuration="31.312481032s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:57:18.66041143 +0000 UTC m=+36.163104324" lastFinishedPulling="2025-05-15 14:57:23.828780726 +0000 UTC m=+41.331473619" observedRunningTime="2025-05-15 14:57:24.307654314 +0000 UTC m=+41.810347217" watchObservedRunningTime="2025-05-15 14:57:24.312481032 +0000 UTC m=+41.815173936" May 15 14:57:25.273938 kubelet[2683]: E0515 14:57:25.273468 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:27.309602 containerd[1559]: time="2025-05-15T14:57:27.309489417Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:27.310815 containerd[1559]: time="2025-05-15T14:57:27.310787251Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=34789138" May 15 14:57:27.311838 containerd[1559]: time="2025-05-15T14:57:27.311772820Z" level=info msg="ImageCreate event name:\"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:27.326118 containerd[1559]: time="2025-05-15T14:57:27.326028508Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:27.327837 containerd[1559]: time="2025-05-15T14:57:27.327693688Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"36281728\" in 3.49801918s" May 15 14:57:27.327837 containerd[1559]: time="2025-05-15T14:57:27.327733268Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" May 15 14:57:27.330173 containerd[1559]: time="2025-05-15T14:57:27.329730864Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 15 14:57:27.354572 containerd[1559]: time="2025-05-15T14:57:27.353897506Z" level=info msg="CreateContainer within sandbox \"935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 15 14:57:27.399206 containerd[1559]: time="2025-05-15T14:57:27.399153235Z" level=info msg="Container f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:27.414719 containerd[1559]: time="2025-05-15T14:57:27.414657672Z" level=info msg="CreateContainer within sandbox \"935f9b9f1004a590dbbb9c84047c2cc8796de655fa241cbfb12bccc5d8a3dccf\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\"" May 15 14:57:27.417025 containerd[1559]: time="2025-05-15T14:57:27.416983801Z" level=info msg="StartContainer for \"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\"" May 15 14:57:27.418888 containerd[1559]: time="2025-05-15T14:57:27.418848420Z" level=info msg="connecting to shim f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a" address="unix:///run/containerd/s/3ff5a9413315124770f10298d98cc053e12a415b2bcf48e53eef81ca70d00104" protocol=ttrpc version=3 May 15 14:57:27.460886 systemd[1]: Started cri-containerd-f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a.scope - libcontainer container f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a. May 15 14:57:27.576432 containerd[1559]: time="2025-05-15T14:57:27.575891735Z" level=info msg="StartContainer for \"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" returns successfully" May 15 14:57:28.182371 containerd[1559]: time="2025-05-15T14:57:28.182032292Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:28.183299 containerd[1559]: time="2025-05-15T14:57:28.183213439Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" May 15 14:57:28.192955 containerd[1559]: time="2025-05-15T14:57:28.192854838Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 862.68909ms" May 15 14:57:28.192955 containerd[1559]: time="2025-05-15T14:57:28.192902001Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 15 14:57:28.196114 containerd[1559]: time="2025-05-15T14:57:28.195937844Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 15 14:57:28.214248 containerd[1559]: time="2025-05-15T14:57:28.213702646Z" level=info msg="CreateContainer within sandbox \"1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 15 14:57:28.236825 containerd[1559]: time="2025-05-15T14:57:28.236780382Z" level=info msg="Container 49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:28.265806 containerd[1559]: time="2025-05-15T14:57:28.265673875Z" level=info msg="CreateContainer within sandbox \"1b9d58e815e115c7439b40dbe191ce573648e3246e0ca2986423369e102f9e1d\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22\"" May 15 14:57:28.267712 containerd[1559]: time="2025-05-15T14:57:28.267643771Z" level=info msg="StartContainer for \"49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22\"" May 15 14:57:28.270376 containerd[1559]: time="2025-05-15T14:57:28.270312913Z" level=info msg="connecting to shim 49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22" address="unix:///run/containerd/s/f4f79168047c2f927091ff7a0a4911abdb77107447b14acbf28aafc3583dc703" protocol=ttrpc version=3 May 15 14:57:28.297213 systemd[1]: Started cri-containerd-49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22.scope - libcontainer container 49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22. May 15 14:57:28.353433 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3636660875.mount: Deactivated successfully. May 15 14:57:28.478506 containerd[1559]: time="2025-05-15T14:57:28.478348772Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"e31d77ebf4893fb917163c3d8eda1e3e508565f4b35472e527a4bcab90563903\" pid:4627 exited_at:{seconds:1747321048 nanos:469866544}" May 15 14:57:28.506564 kubelet[2683]: I0515 14:57:28.506230 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-87fd6d8fb-2ntcg" podStartSLOduration=29.367564558 podStartE2EDuration="35.505935484s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:57:21.190378242 +0000 UTC m=+38.693071124" lastFinishedPulling="2025-05-15 14:57:27.328749156 +0000 UTC m=+44.831442050" observedRunningTime="2025-05-15 14:57:28.37534425 +0000 UTC m=+45.878037152" watchObservedRunningTime="2025-05-15 14:57:28.505935484 +0000 UTC m=+46.008628385" May 15 14:57:28.514592 containerd[1559]: time="2025-05-15T14:57:28.514478648Z" level=info msg="StartContainer for \"49ffac10b703bac494340239bdb2319841632e2eb09ea8fb22c356c50c45aa22\" returns successfully" May 15 14:57:30.342717 kubelet[2683]: I0515 14:57:30.342669 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 14:57:30.422655 systemd[1]: Started sshd@7-24.199.109.202:22-139.178.68.195:55154.service - OpenSSH per-connection server daemon (139.178.68.195:55154). May 15 14:57:30.549663 sshd[4656]: Accepted publickey for core from 139.178.68.195 port 55154 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:30.553285 sshd-session[4656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:30.563636 systemd-logind[1529]: New session 8 of user core. May 15 14:57:30.570037 systemd[1]: Started session-8.scope - Session 8 of User core. May 15 14:57:31.140101 sshd[4658]: Connection closed by 139.178.68.195 port 55154 May 15 14:57:31.141217 sshd-session[4656]: pam_unix(sshd:session): session closed for user core May 15 14:57:31.151222 systemd[1]: sshd@7-24.199.109.202:22-139.178.68.195:55154.service: Deactivated successfully. May 15 14:57:31.156355 systemd[1]: session-8.scope: Deactivated successfully. May 15 14:57:31.160228 systemd-logind[1529]: Session 8 logged out. Waiting for processes to exit. May 15 14:57:31.163207 systemd-logind[1529]: Removed session 8. May 15 14:57:33.094819 containerd[1559]: time="2025-05-15T14:57:33.094731632Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:33.097190 containerd[1559]: time="2025-05-15T14:57:33.097144718Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7912898" May 15 14:57:33.097950 containerd[1559]: time="2025-05-15T14:57:33.097864383Z" level=info msg="ImageCreate event name:\"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:33.100934 containerd[1559]: time="2025-05-15T14:57:33.100588909Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:33.101699 containerd[1559]: time="2025-05-15T14:57:33.101646600Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"9405520\" in 4.905667018s" May 15 14:57:33.101699 containerd[1559]: time="2025-05-15T14:57:33.101681669Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" May 15 14:57:33.109003 containerd[1559]: time="2025-05-15T14:57:33.108941000Z" level=info msg="CreateContainer within sandbox \"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 15 14:57:33.141301 containerd[1559]: time="2025-05-15T14:57:33.139568477Z" level=info msg="Container 886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:33.170346 containerd[1559]: time="2025-05-15T14:57:33.170257078Z" level=info msg="CreateContainer within sandbox \"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07\"" May 15 14:57:33.172282 containerd[1559]: time="2025-05-15T14:57:33.171802843Z" level=info msg="StartContainer for \"886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07\"" May 15 14:57:33.175685 containerd[1559]: time="2025-05-15T14:57:33.175604470Z" level=info msg="connecting to shim 886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07" address="unix:///run/containerd/s/178158c20a30b24b99dfdf5e8426e1a5423b7658001a95f16bc96aa6164049d2" protocol=ttrpc version=3 May 15 14:57:33.215838 systemd[1]: Started cri-containerd-886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07.scope - libcontainer container 886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07. May 15 14:57:33.276465 containerd[1559]: time="2025-05-15T14:57:33.276330252Z" level=info msg="StartContainer for \"886895acb7724965b0b38b38f50a89250cfe0a9d57319fc750ec9c6b2cf29c07\" returns successfully" May 15 14:57:33.281093 containerd[1559]: time="2025-05-15T14:57:33.280972259Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 15 14:57:35.353781 containerd[1559]: time="2025-05-15T14:57:35.353553728Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:35.354750 containerd[1559]: time="2025-05-15T14:57:35.354622538Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13991773" May 15 14:57:35.355769 containerd[1559]: time="2025-05-15T14:57:35.355742906Z" level=info msg="ImageCreate event name:\"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:35.358497 containerd[1559]: time="2025-05-15T14:57:35.358352697Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 15 14:57:35.359504 containerd[1559]: time="2025-05-15T14:57:35.359463857Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"15484347\" in 2.078424596s" May 15 14:57:35.359621 containerd[1559]: time="2025-05-15T14:57:35.359512891Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" May 15 14:57:35.364241 containerd[1559]: time="2025-05-15T14:57:35.364174004Z" level=info msg="CreateContainer within sandbox \"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 15 14:57:35.375548 containerd[1559]: time="2025-05-15T14:57:35.374611896Z" level=info msg="Container a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350: CDI devices from CRI Config.CDIDevices: []" May 15 14:57:35.398346 containerd[1559]: time="2025-05-15T14:57:35.398286567Z" level=info msg="CreateContainer within sandbox \"686ce474524471d867cfabe261f704ee032fa0388bfb96b36b6e515d846180d6\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350\"" May 15 14:57:35.399423 containerd[1559]: time="2025-05-15T14:57:35.399392066Z" level=info msg="StartContainer for \"a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350\"" May 15 14:57:35.401994 containerd[1559]: time="2025-05-15T14:57:35.401958388Z" level=info msg="connecting to shim a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350" address="unix:///run/containerd/s/178158c20a30b24b99dfdf5e8426e1a5423b7658001a95f16bc96aa6164049d2" protocol=ttrpc version=3 May 15 14:57:35.429838 systemd[1]: Started cri-containerd-a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350.scope - libcontainer container a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350. May 15 14:57:35.483200 containerd[1559]: time="2025-05-15T14:57:35.483117362Z" level=info msg="StartContainer for \"a21c517f6983d34a13c5a14dda196802f4f810f3949a2ac003b7b86989a77350\" returns successfully" May 15 14:57:35.990898 kubelet[2683]: I0515 14:57:35.990794 2683 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 15 14:57:35.994677 kubelet[2683]: I0515 14:57:35.994051 2683 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 15 14:57:36.164648 systemd[1]: Started sshd@8-24.199.109.202:22-139.178.68.195:49738.service - OpenSSH per-connection server daemon (139.178.68.195:49738). May 15 14:57:36.292228 sshd[4750]: Accepted publickey for core from 139.178.68.195 port 49738 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:36.294417 sshd-session[4750]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:36.302141 systemd-logind[1529]: New session 9 of user core. May 15 14:57:36.308034 systemd[1]: Started session-9.scope - Session 9 of User core. May 15 14:57:36.395997 kubelet[2683]: I0515 14:57:36.394450 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6684646c96-q9qjj" podStartSLOduration=37.072905031 podStartE2EDuration="43.3944296s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:57:21.874027986 +0000 UTC m=+39.376720868" lastFinishedPulling="2025-05-15 14:57:28.195552556 +0000 UTC m=+45.698245437" observedRunningTime="2025-05-15 14:57:29.360348235 +0000 UTC m=+46.863041136" watchObservedRunningTime="2025-05-15 14:57:36.3944296 +0000 UTC m=+53.897122500" May 15 14:57:36.653864 sshd[4752]: Connection closed by 139.178.68.195 port 49738 May 15 14:57:36.655175 sshd-session[4750]: pam_unix(sshd:session): session closed for user core May 15 14:57:36.661051 systemd[1]: sshd@8-24.199.109.202:22-139.178.68.195:49738.service: Deactivated successfully. May 15 14:57:36.664205 systemd[1]: session-9.scope: Deactivated successfully. May 15 14:57:36.665595 systemd-logind[1529]: Session 9 logged out. Waiting for processes to exit. May 15 14:57:36.668144 systemd-logind[1529]: Removed session 9. May 15 14:57:41.672785 systemd[1]: Started sshd@9-24.199.109.202:22-139.178.68.195:49740.service - OpenSSH per-connection server daemon (139.178.68.195:49740). May 15 14:57:41.747781 sshd[4765]: Accepted publickey for core from 139.178.68.195 port 49740 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:41.750036 sshd-session[4765]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:41.758472 systemd-logind[1529]: New session 10 of user core. May 15 14:57:41.764855 systemd[1]: Started session-10.scope - Session 10 of User core. May 15 14:57:41.930673 sshd[4767]: Connection closed by 139.178.68.195 port 49740 May 15 14:57:41.931730 sshd-session[4765]: pam_unix(sshd:session): session closed for user core May 15 14:57:41.937721 systemd[1]: sshd@9-24.199.109.202:22-139.178.68.195:49740.service: Deactivated successfully. May 15 14:57:41.940466 systemd[1]: session-10.scope: Deactivated successfully. May 15 14:57:41.941509 systemd-logind[1529]: Session 10 logged out. Waiting for processes to exit. May 15 14:57:41.943898 systemd-logind[1529]: Removed session 10. May 15 14:57:45.989729 kubelet[2683]: I0515 14:57:45.989411 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 15 14:57:46.029968 kubelet[2683]: I0515 14:57:46.029852 2683 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-5pnkc" podStartSLOduration=40.552171101 podStartE2EDuration="53.029793725s" podCreationTimestamp="2025-05-15 14:56:53 +0000 UTC" firstStartedPulling="2025-05-15 14:57:22.883784512 +0000 UTC m=+40.386477404" lastFinishedPulling="2025-05-15 14:57:35.361407134 +0000 UTC m=+52.864100028" observedRunningTime="2025-05-15 14:57:36.398185784 +0000 UTC m=+53.900878685" watchObservedRunningTime="2025-05-15 14:57:46.029793725 +0000 UTC m=+63.532486630" May 15 14:57:46.954084 systemd[1]: Started sshd@10-24.199.109.202:22-139.178.68.195:51580.service - OpenSSH per-connection server daemon (139.178.68.195:51580). May 15 14:57:47.039476 sshd[4793]: Accepted publickey for core from 139.178.68.195 port 51580 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:47.042233 sshd-session[4793]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:47.049895 systemd-logind[1529]: New session 11 of user core. May 15 14:57:47.056868 systemd[1]: Started session-11.scope - Session 11 of User core. May 15 14:57:47.244340 sshd[4795]: Connection closed by 139.178.68.195 port 51580 May 15 14:57:47.245693 sshd-session[4793]: pam_unix(sshd:session): session closed for user core May 15 14:57:47.258716 systemd[1]: sshd@10-24.199.109.202:22-139.178.68.195:51580.service: Deactivated successfully. May 15 14:57:47.261995 systemd[1]: session-11.scope: Deactivated successfully. May 15 14:57:47.263917 systemd-logind[1529]: Session 11 logged out. Waiting for processes to exit. May 15 14:57:47.269826 systemd[1]: Started sshd@11-24.199.109.202:22-139.178.68.195:51592.service - OpenSSH per-connection server daemon (139.178.68.195:51592). May 15 14:57:47.271358 systemd-logind[1529]: Removed session 11. May 15 14:57:47.372232 sshd[4807]: Accepted publickey for core from 139.178.68.195 port 51592 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:47.376617 sshd-session[4807]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:47.387766 systemd-logind[1529]: New session 12 of user core. May 15 14:57:47.394086 systemd[1]: Started session-12.scope - Session 12 of User core. May 15 14:57:47.732586 sshd[4809]: Connection closed by 139.178.68.195 port 51592 May 15 14:57:47.734259 sshd-session[4807]: pam_unix(sshd:session): session closed for user core May 15 14:57:47.747308 systemd[1]: sshd@11-24.199.109.202:22-139.178.68.195:51592.service: Deactivated successfully. May 15 14:57:47.753692 systemd[1]: session-12.scope: Deactivated successfully. May 15 14:57:47.758782 systemd-logind[1529]: Session 12 logged out. Waiting for processes to exit. May 15 14:57:47.767927 systemd[1]: Started sshd@12-24.199.109.202:22-139.178.68.195:51602.service - OpenSSH per-connection server daemon (139.178.68.195:51602). May 15 14:57:47.771550 systemd-logind[1529]: Removed session 12. May 15 14:57:47.854552 sshd[4819]: Accepted publickey for core from 139.178.68.195 port 51602 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:47.858480 sshd-session[4819]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:47.870065 systemd-logind[1529]: New session 13 of user core. May 15 14:57:47.875065 systemd[1]: Started session-13.scope - Session 13 of User core. May 15 14:57:48.133707 sshd[4821]: Connection closed by 139.178.68.195 port 51602 May 15 14:57:48.134208 sshd-session[4819]: pam_unix(sshd:session): session closed for user core May 15 14:57:48.144449 systemd[1]: sshd@12-24.199.109.202:22-139.178.68.195:51602.service: Deactivated successfully. May 15 14:57:48.153385 systemd[1]: session-13.scope: Deactivated successfully. May 15 14:57:48.162608 systemd-logind[1529]: Session 13 logged out. Waiting for processes to exit. May 15 14:57:48.164780 systemd-logind[1529]: Removed session 13. May 15 14:57:48.310502 containerd[1559]: time="2025-05-15T14:57:48.310438722Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"901c73723e71db66c870271ec8eee11327a1ad96ec0c70d95debd94b67a771e5\" pid:4844 exited_at:{seconds:1747321068 nanos:308839052}" May 15 14:57:48.320543 kubelet[2683]: E0515 14:57:48.320431 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:53.154235 systemd[1]: Started sshd@13-24.199.109.202:22-139.178.68.195:51616.service - OpenSSH per-connection server daemon (139.178.68.195:51616). May 15 14:57:53.224902 sshd[4868]: Accepted publickey for core from 139.178.68.195 port 51616 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:53.231679 sshd-session[4868]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:53.238311 systemd-logind[1529]: New session 14 of user core. May 15 14:57:53.243792 systemd[1]: Started session-14.scope - Session 14 of User core. May 15 14:57:53.497871 sshd[4870]: Connection closed by 139.178.68.195 port 51616 May 15 14:57:53.498651 sshd-session[4868]: pam_unix(sshd:session): session closed for user core May 15 14:57:53.505636 systemd[1]: sshd@13-24.199.109.202:22-139.178.68.195:51616.service: Deactivated successfully. May 15 14:57:53.508512 systemd[1]: session-14.scope: Deactivated successfully. May 15 14:57:53.509867 systemd-logind[1529]: Session 14 logged out. Waiting for processes to exit. May 15 14:57:53.512262 systemd-logind[1529]: Removed session 14. May 15 14:57:58.416561 containerd[1559]: time="2025-05-15T14:57:58.416466018Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"c3ef3c4cd25213281ba22ab83dd447c2a81b3755ddb246f67ebd5300d58d081e\" pid:4894 exited_at:{seconds:1747321078 nanos:415467777}" May 15 14:57:58.521143 systemd[1]: Started sshd@14-24.199.109.202:22-139.178.68.195:47996.service - OpenSSH per-connection server daemon (139.178.68.195:47996). May 15 14:57:58.622307 sshd[4905]: Accepted publickey for core from 139.178.68.195 port 47996 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:57:58.625447 sshd-session[4905]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:57:58.633952 systemd-logind[1529]: New session 15 of user core. May 15 14:57:58.648871 systemd[1]: Started session-15.scope - Session 15 of User core. May 15 14:57:58.710672 kubelet[2683]: E0515 14:57:58.710345 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:57:58.854187 sshd[4907]: Connection closed by 139.178.68.195 port 47996 May 15 14:57:58.854774 sshd-session[4905]: pam_unix(sshd:session): session closed for user core May 15 14:57:58.860481 systemd[1]: sshd@14-24.199.109.202:22-139.178.68.195:47996.service: Deactivated successfully. May 15 14:57:58.863384 systemd[1]: session-15.scope: Deactivated successfully. May 15 14:57:58.865583 systemd-logind[1529]: Session 15 logged out. Waiting for processes to exit. May 15 14:57:58.867424 systemd-logind[1529]: Removed session 15. May 15 14:58:03.877335 systemd[1]: Started sshd@15-24.199.109.202:22-139.178.68.195:53360.service - OpenSSH per-connection server daemon (139.178.68.195:53360). May 15 14:58:03.956480 sshd[4926]: Accepted publickey for core from 139.178.68.195 port 53360 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:03.958925 sshd-session[4926]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:03.967464 systemd-logind[1529]: New session 16 of user core. May 15 14:58:03.973879 systemd[1]: Started session-16.scope - Session 16 of User core. May 15 14:58:04.182497 sshd[4928]: Connection closed by 139.178.68.195 port 53360 May 15 14:58:04.183950 sshd-session[4926]: pam_unix(sshd:session): session closed for user core May 15 14:58:04.188919 systemd[1]: sshd@15-24.199.109.202:22-139.178.68.195:53360.service: Deactivated successfully. May 15 14:58:04.193447 systemd[1]: session-16.scope: Deactivated successfully. May 15 14:58:04.197072 systemd-logind[1529]: Session 16 logged out. Waiting for processes to exit. May 15 14:58:04.199303 systemd-logind[1529]: Removed session 16. May 15 14:58:08.710820 kubelet[2683]: E0515 14:58:08.710254 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:09.205491 systemd[1]: Started sshd@16-24.199.109.202:22-139.178.68.195:53364.service - OpenSSH per-connection server daemon (139.178.68.195:53364). May 15 14:58:09.271131 sshd[4941]: Accepted publickey for core from 139.178.68.195 port 53364 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:09.273544 sshd-session[4941]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:09.281792 systemd-logind[1529]: New session 17 of user core. May 15 14:58:09.286810 systemd[1]: Started session-17.scope - Session 17 of User core. May 15 14:58:09.456625 sshd[4943]: Connection closed by 139.178.68.195 port 53364 May 15 14:58:09.457334 sshd-session[4941]: pam_unix(sshd:session): session closed for user core May 15 14:58:09.473103 systemd[1]: sshd@16-24.199.109.202:22-139.178.68.195:53364.service: Deactivated successfully. May 15 14:58:09.477359 systemd[1]: session-17.scope: Deactivated successfully. May 15 14:58:09.479749 systemd-logind[1529]: Session 17 logged out. Waiting for processes to exit. May 15 14:58:09.486881 systemd[1]: Started sshd@17-24.199.109.202:22-139.178.68.195:53374.service - OpenSSH per-connection server daemon (139.178.68.195:53374). May 15 14:58:09.488966 systemd-logind[1529]: Removed session 17. May 15 14:58:09.567295 sshd[4954]: Accepted publickey for core from 139.178.68.195 port 53374 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:09.569303 sshd-session[4954]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:09.578408 systemd-logind[1529]: New session 18 of user core. May 15 14:58:09.583192 systemd[1]: Started session-18.scope - Session 18 of User core. May 15 14:58:09.963729 sshd[4956]: Connection closed by 139.178.68.195 port 53374 May 15 14:58:09.965284 sshd-session[4954]: pam_unix(sshd:session): session closed for user core May 15 14:58:09.979335 systemd[1]: sshd@17-24.199.109.202:22-139.178.68.195:53374.service: Deactivated successfully. May 15 14:58:09.983210 systemd[1]: session-18.scope: Deactivated successfully. May 15 14:58:09.986934 systemd-logind[1529]: Session 18 logged out. Waiting for processes to exit. May 15 14:58:09.991717 systemd[1]: Started sshd@18-24.199.109.202:22-139.178.68.195:53378.service - OpenSSH per-connection server daemon (139.178.68.195:53378). May 15 14:58:09.993895 systemd-logind[1529]: Removed session 18. May 15 14:58:10.122563 sshd[4966]: Accepted publickey for core from 139.178.68.195 port 53378 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:10.123890 sshd-session[4966]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:10.131645 systemd-logind[1529]: New session 19 of user core. May 15 14:58:10.140842 systemd[1]: Started session-19.scope - Session 19 of User core. May 15 14:58:11.294065 sshd[4968]: Connection closed by 139.178.68.195 port 53378 May 15 14:58:11.294521 sshd-session[4966]: pam_unix(sshd:session): session closed for user core May 15 14:58:11.312581 systemd[1]: sshd@18-24.199.109.202:22-139.178.68.195:53378.service: Deactivated successfully. May 15 14:58:11.317277 systemd[1]: session-19.scope: Deactivated successfully. May 15 14:58:11.320797 systemd-logind[1529]: Session 19 logged out. Waiting for processes to exit. May 15 14:58:11.327837 systemd-logind[1529]: Removed session 19. May 15 14:58:11.333594 systemd[1]: Started sshd@19-24.199.109.202:22-139.178.68.195:53394.service - OpenSSH per-connection server daemon (139.178.68.195:53394). May 15 14:58:11.431549 sshd[4982]: Accepted publickey for core from 139.178.68.195 port 53394 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:11.433629 sshd-session[4982]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:11.443632 systemd-logind[1529]: New session 20 of user core. May 15 14:58:11.450226 systemd[1]: Started session-20.scope - Session 20 of User core. May 15 14:58:12.068591 sshd[4984]: Connection closed by 139.178.68.195 port 53394 May 15 14:58:12.069786 sshd-session[4982]: pam_unix(sshd:session): session closed for user core May 15 14:58:12.085818 systemd[1]: sshd@19-24.199.109.202:22-139.178.68.195:53394.service: Deactivated successfully. May 15 14:58:12.089251 systemd[1]: session-20.scope: Deactivated successfully. May 15 14:58:12.091513 systemd-logind[1529]: Session 20 logged out. Waiting for processes to exit. May 15 14:58:12.095929 systemd-logind[1529]: Removed session 20. May 15 14:58:12.099199 systemd[1]: Started sshd@20-24.199.109.202:22-139.178.68.195:53396.service - OpenSSH per-connection server daemon (139.178.68.195:53396). May 15 14:58:12.171820 sshd[4997]: Accepted publickey for core from 139.178.68.195 port 53396 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:12.174140 sshd-session[4997]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:12.181671 systemd-logind[1529]: New session 21 of user core. May 15 14:58:12.186925 systemd[1]: Started session-21.scope - Session 21 of User core. May 15 14:58:12.333835 sshd[4999]: Connection closed by 139.178.68.195 port 53396 May 15 14:58:12.335002 sshd-session[4997]: pam_unix(sshd:session): session closed for user core May 15 14:58:12.341811 systemd[1]: sshd@20-24.199.109.202:22-139.178.68.195:53396.service: Deactivated successfully. May 15 14:58:12.344835 systemd[1]: session-21.scope: Deactivated successfully. May 15 14:58:12.346582 systemd-logind[1529]: Session 21 logged out. Waiting for processes to exit. May 15 14:58:12.349493 systemd-logind[1529]: Removed session 21. May 15 14:58:15.709743 kubelet[2683]: E0515 14:58:15.709615 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:17.357740 systemd[1]: Started sshd@21-24.199.109.202:22-139.178.68.195:52366.service - OpenSSH per-connection server daemon (139.178.68.195:52366). May 15 14:58:17.432869 sshd[5011]: Accepted publickey for core from 139.178.68.195 port 52366 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:17.435755 sshd-session[5011]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:17.445251 systemd-logind[1529]: New session 22 of user core. May 15 14:58:17.454864 systemd[1]: Started session-22.scope - Session 22 of User core. May 15 14:58:17.769277 sshd[5013]: Connection closed by 139.178.68.195 port 52366 May 15 14:58:17.770859 sshd-session[5011]: pam_unix(sshd:session): session closed for user core May 15 14:58:17.784487 systemd[1]: sshd@21-24.199.109.202:22-139.178.68.195:52366.service: Deactivated successfully. May 15 14:58:17.787670 systemd[1]: session-22.scope: Deactivated successfully. May 15 14:58:17.790889 systemd-logind[1529]: Session 22 logged out. Waiting for processes to exit. May 15 14:58:17.792565 systemd-logind[1529]: Removed session 22. May 15 14:58:18.194668 containerd[1559]: time="2025-05-15T14:58:18.194103093Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"c60b6cfe6eaa06db83d28f4b948934576e1b353bde92ce39887f8034ba9f7557\" pid:5038 exited_at:{seconds:1747321098 nanos:193457381}" May 15 14:58:22.787619 systemd[1]: Started sshd@22-24.199.109.202:22-139.178.68.195:52368.service - OpenSSH per-connection server daemon (139.178.68.195:52368). May 15 14:58:22.898202 sshd[5051]: Accepted publickey for core from 139.178.68.195 port 52368 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:22.900485 sshd-session[5051]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:22.909741 systemd-logind[1529]: New session 23 of user core. May 15 14:58:22.915812 systemd[1]: Started session-23.scope - Session 23 of User core. May 15 14:58:23.111775 sshd[5053]: Connection closed by 139.178.68.195 port 52368 May 15 14:58:23.112919 sshd-session[5051]: pam_unix(sshd:session): session closed for user core May 15 14:58:23.121501 systemd[1]: sshd@22-24.199.109.202:22-139.178.68.195:52368.service: Deactivated successfully. May 15 14:58:23.127455 systemd[1]: session-23.scope: Deactivated successfully. May 15 14:58:23.130641 systemd-logind[1529]: Session 23 logged out. Waiting for processes to exit. May 15 14:58:23.136648 systemd-logind[1529]: Removed session 23. May 15 14:58:24.713855 kubelet[2683]: E0515 14:58:24.713692 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:28.132372 systemd[1]: Started sshd@23-24.199.109.202:22-139.178.68.195:39464.service - OpenSSH per-connection server daemon (139.178.68.195:39464). May 15 14:58:28.273054 sshd[5065]: Accepted publickey for core from 139.178.68.195 port 39464 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:28.276178 sshd-session[5065]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:28.284612 systemd-logind[1529]: New session 24 of user core. May 15 14:58:28.289829 systemd[1]: Started session-24.scope - Session 24 of User core. May 15 14:58:28.451191 containerd[1559]: time="2025-05-15T14:58:28.451046403Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"a402a1b05f789a4c84ddfd4c2eae157dd7820b89291d15887123c009c4240b67\" pid:5081 exited_at:{seconds:1747321108 nanos:450286605}" May 15 14:58:28.490868 containerd[1559]: time="2025-05-15T14:58:28.490809439Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"dbd687184465a513b04dc5df03900381180bb99198701ee1be38a405741fa06b\" pid:5100 exited_at:{seconds:1747321108 nanos:490137978}" May 15 14:58:28.767232 sshd[5067]: Connection closed by 139.178.68.195 port 39464 May 15 14:58:28.767758 sshd-session[5065]: pam_unix(sshd:session): session closed for user core May 15 14:58:28.775147 systemd-logind[1529]: Session 24 logged out. Waiting for processes to exit. May 15 14:58:28.775228 systemd[1]: sshd@23-24.199.109.202:22-139.178.68.195:39464.service: Deactivated successfully. May 15 14:58:28.780264 systemd[1]: session-24.scope: Deactivated successfully. May 15 14:58:28.786439 systemd-logind[1529]: Removed session 24. May 15 14:58:31.709862 kubelet[2683]: E0515 14:58:31.709820 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:33.783037 systemd[1]: Started sshd@24-24.199.109.202:22-139.178.68.195:60480.service - OpenSSH per-connection server daemon (139.178.68.195:60480). May 15 14:58:33.878821 sshd[5120]: Accepted publickey for core from 139.178.68.195 port 60480 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:33.881436 sshd-session[5120]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:33.896271 systemd-logind[1529]: New session 25 of user core. May 15 14:58:33.899833 systemd[1]: Started session-25.scope - Session 25 of User core. May 15 14:58:34.078637 sshd[5122]: Connection closed by 139.178.68.195 port 60480 May 15 14:58:34.078884 sshd-session[5120]: pam_unix(sshd:session): session closed for user core May 15 14:58:34.084962 systemd[1]: sshd@24-24.199.109.202:22-139.178.68.195:60480.service: Deactivated successfully. May 15 14:58:34.088586 systemd[1]: session-25.scope: Deactivated successfully. May 15 14:58:34.090635 systemd-logind[1529]: Session 25 logged out. Waiting for processes to exit. May 15 14:58:34.094642 systemd-logind[1529]: Removed session 25. May 15 14:58:34.710548 kubelet[2683]: E0515 14:58:34.709878 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:39.098971 systemd[1]: Started sshd@25-24.199.109.202:22-139.178.68.195:60496.service - OpenSSH per-connection server daemon (139.178.68.195:60496). May 15 14:58:39.176117 sshd[5134]: Accepted publickey for core from 139.178.68.195 port 60496 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:39.178355 sshd-session[5134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:39.186680 systemd-logind[1529]: New session 26 of user core. May 15 14:58:39.191836 systemd[1]: Started session-26.scope - Session 26 of User core. May 15 14:58:39.341760 sshd[5136]: Connection closed by 139.178.68.195 port 60496 May 15 14:58:39.342814 sshd-session[5134]: pam_unix(sshd:session): session closed for user core May 15 14:58:39.348650 systemd[1]: sshd@25-24.199.109.202:22-139.178.68.195:60496.service: Deactivated successfully. May 15 14:58:39.352367 systemd[1]: session-26.scope: Deactivated successfully. May 15 14:58:39.354838 systemd-logind[1529]: Session 26 logged out. Waiting for processes to exit. May 15 14:58:39.358592 systemd-logind[1529]: Removed session 26. May 15 14:58:44.363860 systemd[1]: Started sshd@26-24.199.109.202:22-139.178.68.195:42864.service - OpenSSH per-connection server daemon (139.178.68.195:42864). May 15 14:58:44.440725 sshd[5157]: Accepted publickey for core from 139.178.68.195 port 42864 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:44.443138 sshd-session[5157]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:44.452241 systemd-logind[1529]: New session 27 of user core. May 15 14:58:44.457929 systemd[1]: Started session-27.scope - Session 27 of User core. May 15 14:58:44.631682 sshd[5159]: Connection closed by 139.178.68.195 port 42864 May 15 14:58:44.632881 sshd-session[5157]: pam_unix(sshd:session): session closed for user core May 15 14:58:44.639152 systemd[1]: sshd@26-24.199.109.202:22-139.178.68.195:42864.service: Deactivated successfully. May 15 14:58:44.643493 systemd[1]: session-27.scope: Deactivated successfully. May 15 14:58:44.645256 systemd-logind[1529]: Session 27 logged out. Waiting for processes to exit. May 15 14:58:44.648430 systemd-logind[1529]: Removed session 27. May 15 14:58:48.175906 containerd[1559]: time="2025-05-15T14:58:48.175834252Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"f29a2249b2ada8acdd5fec6e6b9eb848b5c89aef5446a5191b60134fe24019c2\" pid:5185 exited_at:{seconds:1747321128 nanos:175190938}" May 15 14:58:49.648138 systemd[1]: Started sshd@27-24.199.109.202:22-139.178.68.195:42876.service - OpenSSH per-connection server daemon (139.178.68.195:42876). May 15 14:58:49.743738 sshd[5197]: Accepted publickey for core from 139.178.68.195 port 42876 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:49.746467 sshd-session[5197]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:49.755618 systemd-logind[1529]: New session 28 of user core. May 15 14:58:49.763814 systemd[1]: Started session-28.scope - Session 28 of User core. May 15 14:58:49.993729 sshd[5199]: Connection closed by 139.178.68.195 port 42876 May 15 14:58:49.994749 sshd-session[5197]: pam_unix(sshd:session): session closed for user core May 15 14:58:50.001558 systemd[1]: sshd@27-24.199.109.202:22-139.178.68.195:42876.service: Deactivated successfully. May 15 14:58:50.005063 systemd[1]: session-28.scope: Deactivated successfully. May 15 14:58:50.007127 systemd-logind[1529]: Session 28 logged out. Waiting for processes to exit. May 15 14:58:50.010359 systemd-logind[1529]: Removed session 28. May 15 14:58:53.710563 kubelet[2683]: E0515 14:58:53.710283 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:58:55.010846 systemd[1]: Started sshd@28-24.199.109.202:22-139.178.68.195:39078.service - OpenSSH per-connection server daemon (139.178.68.195:39078). May 15 14:58:55.085550 sshd[5217]: Accepted publickey for core from 139.178.68.195 port 39078 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:58:55.086953 sshd-session[5217]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:58:55.100628 systemd-logind[1529]: New session 29 of user core. May 15 14:58:55.106309 systemd[1]: Started session-29.scope - Session 29 of User core. May 15 14:58:55.321496 sshd[5219]: Connection closed by 139.178.68.195 port 39078 May 15 14:58:55.321869 sshd-session[5217]: pam_unix(sshd:session): session closed for user core May 15 14:58:55.330259 systemd[1]: sshd@28-24.199.109.202:22-139.178.68.195:39078.service: Deactivated successfully. May 15 14:58:55.336329 systemd[1]: session-29.scope: Deactivated successfully. May 15 14:58:55.340262 systemd-logind[1529]: Session 29 logged out. Waiting for processes to exit. May 15 14:58:55.343738 systemd-logind[1529]: Removed session 29. May 15 14:58:58.385315 containerd[1559]: time="2025-05-15T14:58:58.385270654Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"32027c61a35422cb95f335b17fba073a38bad2bec5ab42d4524bb0e4c2471681\" pid:5254 exited_at:{seconds:1747321138 nanos:384611841}" May 15 14:59:00.343946 systemd[1]: Started sshd@29-24.199.109.202:22-139.178.68.195:39094.service - OpenSSH per-connection server daemon (139.178.68.195:39094). May 15 14:59:00.423180 sshd[5264]: Accepted publickey for core from 139.178.68.195 port 39094 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:00.425633 sshd-session[5264]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:00.431973 systemd-logind[1529]: New session 30 of user core. May 15 14:59:00.440942 systemd[1]: Started session-30.scope - Session 30 of User core. May 15 14:59:00.665339 sshd[5266]: Connection closed by 139.178.68.195 port 39094 May 15 14:59:00.666802 sshd-session[5264]: pam_unix(sshd:session): session closed for user core May 15 14:59:00.671381 systemd-logind[1529]: Session 30 logged out. Waiting for processes to exit. May 15 14:59:00.671632 systemd[1]: sshd@29-24.199.109.202:22-139.178.68.195:39094.service: Deactivated successfully. May 15 14:59:00.675182 systemd[1]: session-30.scope: Deactivated successfully. May 15 14:59:00.679659 systemd-logind[1529]: Removed session 30. May 15 14:59:05.684302 systemd[1]: Started sshd@30-24.199.109.202:22-139.178.68.195:33292.service - OpenSSH per-connection server daemon (139.178.68.195:33292). May 15 14:59:05.817212 sshd[5278]: Accepted publickey for core from 139.178.68.195 port 33292 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:05.819123 sshd-session[5278]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:05.828540 systemd-logind[1529]: New session 31 of user core. May 15 14:59:05.833841 systemd[1]: Started session-31.scope - Session 31 of User core. May 15 14:59:06.055855 sshd[5280]: Connection closed by 139.178.68.195 port 33292 May 15 14:59:06.056315 sshd-session[5278]: pam_unix(sshd:session): session closed for user core May 15 14:59:06.063073 systemd[1]: sshd@30-24.199.109.202:22-139.178.68.195:33292.service: Deactivated successfully. May 15 14:59:06.067401 systemd[1]: session-31.scope: Deactivated successfully. May 15 14:59:06.068981 systemd-logind[1529]: Session 31 logged out. Waiting for processes to exit. May 15 14:59:06.071293 systemd-logind[1529]: Removed session 31. May 15 14:59:06.711361 kubelet[2683]: E0515 14:59:06.709638 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:10.710674 kubelet[2683]: E0515 14:59:10.710404 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:11.074371 systemd[1]: Started sshd@31-24.199.109.202:22-139.178.68.195:33302.service - OpenSSH per-connection server daemon (139.178.68.195:33302). May 15 14:59:11.150760 sshd[5291]: Accepted publickey for core from 139.178.68.195 port 33302 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:11.153516 sshd-session[5291]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:11.163184 systemd-logind[1529]: New session 32 of user core. May 15 14:59:11.174846 systemd[1]: Started session-32.scope - Session 32 of User core. May 15 14:59:11.353401 sshd[5293]: Connection closed by 139.178.68.195 port 33302 May 15 14:59:11.354313 sshd-session[5291]: pam_unix(sshd:session): session closed for user core May 15 14:59:11.361282 systemd[1]: sshd@31-24.199.109.202:22-139.178.68.195:33302.service: Deactivated successfully. May 15 14:59:11.364243 systemd[1]: session-32.scope: Deactivated successfully. May 15 14:59:11.366270 systemd-logind[1529]: Session 32 logged out. Waiting for processes to exit. May 15 14:59:11.369718 systemd-logind[1529]: Removed session 32. May 15 14:59:16.372980 systemd[1]: Started sshd@32-24.199.109.202:22-139.178.68.195:50482.service - OpenSSH per-connection server daemon (139.178.68.195:50482). May 15 14:59:16.466123 sshd[5305]: Accepted publickey for core from 139.178.68.195 port 50482 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:16.468907 sshd-session[5305]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:16.478967 systemd-logind[1529]: New session 33 of user core. May 15 14:59:16.487283 systemd[1]: Started session-33.scope - Session 33 of User core. May 15 14:59:16.704209 sshd[5307]: Connection closed by 139.178.68.195 port 50482 May 15 14:59:16.705292 sshd-session[5305]: pam_unix(sshd:session): session closed for user core May 15 14:59:16.710996 systemd-logind[1529]: Session 33 logged out. Waiting for processes to exit. May 15 14:59:16.711950 systemd[1]: sshd@32-24.199.109.202:22-139.178.68.195:50482.service: Deactivated successfully. May 15 14:59:16.716116 systemd[1]: session-33.scope: Deactivated successfully. May 15 14:59:16.721090 systemd-logind[1529]: Removed session 33. May 15 14:59:18.188094 containerd[1559]: time="2025-05-15T14:59:18.187816147Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"72fa76ed08f8f2aee001026e848cdf3d696562698546f7dd1d70b383cb0e185b\" pid:5332 exited_at:{seconds:1747321158 nanos:186924905}" May 15 14:59:21.719474 systemd[1]: Started sshd@33-24.199.109.202:22-139.178.68.195:50498.service - OpenSSH per-connection server daemon (139.178.68.195:50498). May 15 14:59:21.850439 sshd[5345]: Accepted publickey for core from 139.178.68.195 port 50498 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:21.853143 sshd-session[5345]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:21.860507 systemd-logind[1529]: New session 34 of user core. May 15 14:59:21.865789 systemd[1]: Started session-34.scope - Session 34 of User core. May 15 14:59:22.284621 sshd[5347]: Connection closed by 139.178.68.195 port 50498 May 15 14:59:22.285152 sshd-session[5345]: pam_unix(sshd:session): session closed for user core May 15 14:59:22.292160 systemd[1]: sshd@33-24.199.109.202:22-139.178.68.195:50498.service: Deactivated successfully. May 15 14:59:22.296645 systemd[1]: session-34.scope: Deactivated successfully. May 15 14:59:22.298662 systemd-logind[1529]: Session 34 logged out. Waiting for processes to exit. May 15 14:59:22.301359 systemd-logind[1529]: Removed session 34. May 15 14:59:24.710184 kubelet[2683]: E0515 14:59:24.710107 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:27.302289 systemd[1]: Started sshd@34-24.199.109.202:22-139.178.68.195:41480.service - OpenSSH per-connection server daemon (139.178.68.195:41480). May 15 14:59:27.378362 sshd[5358]: Accepted publickey for core from 139.178.68.195 port 41480 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:27.380786 sshd-session[5358]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:27.388159 systemd-logind[1529]: New session 35 of user core. May 15 14:59:27.396920 systemd[1]: Started session-35.scope - Session 35 of User core. May 15 14:59:27.561654 sshd[5360]: Connection closed by 139.178.68.195 port 41480 May 15 14:59:27.562702 sshd-session[5358]: pam_unix(sshd:session): session closed for user core May 15 14:59:27.569219 systemd[1]: sshd@34-24.199.109.202:22-139.178.68.195:41480.service: Deactivated successfully. May 15 14:59:27.573889 systemd[1]: session-35.scope: Deactivated successfully. May 15 14:59:27.576603 systemd-logind[1529]: Session 35 logged out. Waiting for processes to exit. May 15 14:59:27.578376 systemd-logind[1529]: Removed session 35. May 15 14:59:28.377701 containerd[1559]: time="2025-05-15T14:59:28.377651644Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"fd81d49e9c897f6a3aa40b65efa2cb0ea7329c05996748de51daf5000a6f7ae4\" pid:5382 exited_at:{seconds:1747321168 nanos:377063284}" May 15 14:59:28.412410 containerd[1559]: time="2025-05-15T14:59:28.412331799Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"d7ae5409cc5d3d5871604d065458f8901ffce270d65fb70ce1b489ccdb19d726\" pid:5401 exited_at:{seconds:1747321168 nanos:411369927}" May 15 14:59:28.709669 kubelet[2683]: E0515 14:59:28.709399 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:32.581503 systemd[1]: Started sshd@35-24.199.109.202:22-139.178.68.195:41496.service - OpenSSH per-connection server daemon (139.178.68.195:41496). May 15 14:59:32.665825 sshd[5415]: Accepted publickey for core from 139.178.68.195 port 41496 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:32.669394 sshd-session[5415]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:32.677036 systemd-logind[1529]: New session 36 of user core. May 15 14:59:32.688952 systemd[1]: Started session-36.scope - Session 36 of User core. May 15 14:59:32.962400 sshd[5417]: Connection closed by 139.178.68.195 port 41496 May 15 14:59:32.963472 sshd-session[5415]: pam_unix(sshd:session): session closed for user core May 15 14:59:32.971424 systemd[1]: sshd@35-24.199.109.202:22-139.178.68.195:41496.service: Deactivated successfully. May 15 14:59:32.975625 systemd[1]: session-36.scope: Deactivated successfully. May 15 14:59:32.977696 systemd-logind[1529]: Session 36 logged out. Waiting for processes to exit. May 15 14:59:32.980627 systemd-logind[1529]: Removed session 36. May 15 14:59:37.984308 systemd[1]: Started sshd@36-24.199.109.202:22-139.178.68.195:40822.service - OpenSSH per-connection server daemon (139.178.68.195:40822). May 15 14:59:38.053536 sshd[5429]: Accepted publickey for core from 139.178.68.195 port 40822 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:38.056155 sshd-session[5429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:38.064814 systemd-logind[1529]: New session 37 of user core. May 15 14:59:38.071879 systemd[1]: Started session-37.scope - Session 37 of User core. May 15 14:59:38.240252 sshd[5431]: Connection closed by 139.178.68.195 port 40822 May 15 14:59:38.241195 sshd-session[5429]: pam_unix(sshd:session): session closed for user core May 15 14:59:38.247088 systemd[1]: sshd@36-24.199.109.202:22-139.178.68.195:40822.service: Deactivated successfully. May 15 14:59:38.249517 systemd[1]: session-37.scope: Deactivated successfully. May 15 14:59:38.253553 systemd-logind[1529]: Session 37 logged out. Waiting for processes to exit. May 15 14:59:38.255407 systemd-logind[1529]: Removed session 37. May 15 14:59:43.263891 systemd[1]: Started sshd@37-24.199.109.202:22-139.178.68.195:40838.service - OpenSSH per-connection server daemon (139.178.68.195:40838). May 15 14:59:43.334548 sshd[5444]: Accepted publickey for core from 139.178.68.195 port 40838 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:43.336296 sshd-session[5444]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:43.343422 systemd-logind[1529]: New session 38 of user core. May 15 14:59:43.351823 systemd[1]: Started session-38.scope - Session 38 of User core. May 15 14:59:43.498707 sshd[5446]: Connection closed by 139.178.68.195 port 40838 May 15 14:59:43.499342 sshd-session[5444]: pam_unix(sshd:session): session closed for user core May 15 14:59:43.505622 systemd[1]: sshd@37-24.199.109.202:22-139.178.68.195:40838.service: Deactivated successfully. May 15 14:59:43.509236 systemd[1]: session-38.scope: Deactivated successfully. May 15 14:59:43.510685 systemd-logind[1529]: Session 38 logged out. Waiting for processes to exit. May 15 14:59:43.513288 systemd-logind[1529]: Removed session 38. May 15 14:59:44.710538 kubelet[2683]: E0515 14:59:44.709917 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:48.202988 containerd[1559]: time="2025-05-15T14:59:48.202839971Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"0755d90fb4b590a44ea925c8772e8fd130b7b175eef1dee1290aaf247d56de56\" pid:5471 exited_at:{seconds:1747321188 nanos:202418794}" May 15 14:59:48.520094 systemd[1]: Started sshd@38-24.199.109.202:22-139.178.68.195:50520.service - OpenSSH per-connection server daemon (139.178.68.195:50520). May 15 14:59:48.605146 sshd[5485]: Accepted publickey for core from 139.178.68.195 port 50520 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:48.608405 sshd-session[5485]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:48.619554 systemd-logind[1529]: New session 39 of user core. May 15 14:59:48.623748 systemd[1]: Started session-39.scope - Session 39 of User core. May 15 14:59:48.795349 sshd[5489]: Connection closed by 139.178.68.195 port 50520 May 15 14:59:48.796272 sshd-session[5485]: pam_unix(sshd:session): session closed for user core May 15 14:59:48.802264 systemd-logind[1529]: Session 39 logged out. Waiting for processes to exit. May 15 14:59:48.802400 systemd[1]: sshd@38-24.199.109.202:22-139.178.68.195:50520.service: Deactivated successfully. May 15 14:59:48.806249 systemd[1]: session-39.scope: Deactivated successfully. May 15 14:59:48.809291 systemd-logind[1529]: Removed session 39. May 15 14:59:52.712719 kubelet[2683]: E0515 14:59:52.712230 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:53.815955 systemd[1]: Started sshd@39-24.199.109.202:22-139.178.68.195:44708.service - OpenSSH per-connection server daemon (139.178.68.195:44708). May 15 14:59:53.894275 sshd[5502]: Accepted publickey for core from 139.178.68.195 port 44708 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:53.896827 sshd-session[5502]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:53.911613 systemd-logind[1529]: New session 40 of user core. May 15 14:59:53.915846 systemd[1]: Started session-40.scope - Session 40 of User core. May 15 14:59:54.094578 sshd[5505]: Connection closed by 139.178.68.195 port 44708 May 15 14:59:54.095648 sshd-session[5502]: pam_unix(sshd:session): session closed for user core May 15 14:59:54.103143 systemd-logind[1529]: Session 40 logged out. Waiting for processes to exit. May 15 14:59:54.103921 systemd[1]: sshd@39-24.199.109.202:22-139.178.68.195:44708.service: Deactivated successfully. May 15 14:59:54.107945 systemd[1]: session-40.scope: Deactivated successfully. May 15 14:59:54.111638 systemd-logind[1529]: Removed session 40. May 15 14:59:57.710407 kubelet[2683]: E0515 14:59:57.710321 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 14:59:58.386916 containerd[1559]: time="2025-05-15T14:59:58.386850628Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"1d38fc5d1d916c9ddaf3740418762d395586201aba43483622f5b287adb68a80\" pid:5528 exited_at:{seconds:1747321198 nanos:385511234}" May 15 14:59:59.118409 systemd[1]: Started sshd@40-24.199.109.202:22-139.178.68.195:44722.service - OpenSSH per-connection server daemon (139.178.68.195:44722). May 15 14:59:59.200047 sshd[5539]: Accepted publickey for core from 139.178.68.195 port 44722 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 14:59:59.202076 sshd-session[5539]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 14:59:59.208600 systemd-logind[1529]: New session 41 of user core. May 15 14:59:59.214785 systemd[1]: Started session-41.scope - Session 41 of User core. May 15 14:59:59.354478 sshd[5541]: Connection closed by 139.178.68.195 port 44722 May 15 14:59:59.355342 sshd-session[5539]: pam_unix(sshd:session): session closed for user core May 15 14:59:59.359742 systemd-logind[1529]: Session 41 logged out. Waiting for processes to exit. May 15 14:59:59.359977 systemd[1]: sshd@40-24.199.109.202:22-139.178.68.195:44722.service: Deactivated successfully. May 15 14:59:59.362762 systemd[1]: session-41.scope: Deactivated successfully. May 15 14:59:59.366400 systemd-logind[1529]: Removed session 41. May 15 15:00:04.392010 systemd[1]: Started sshd@41-24.199.109.202:22-139.178.68.195:49050.service - OpenSSH per-connection server daemon (139.178.68.195:49050). May 15 15:00:04.491093 sshd[5561]: Accepted publickey for core from 139.178.68.195 port 49050 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:04.494953 sshd-session[5561]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:04.511629 systemd-logind[1529]: New session 42 of user core. May 15 15:00:04.521187 systemd[1]: Started session-42.scope - Session 42 of User core. May 15 15:00:04.765443 sshd[5563]: Connection closed by 139.178.68.195 port 49050 May 15 15:00:04.765209 sshd-session[5561]: pam_unix(sshd:session): session closed for user core May 15 15:00:04.772997 systemd[1]: sshd@41-24.199.109.202:22-139.178.68.195:49050.service: Deactivated successfully. May 15 15:00:04.779711 systemd[1]: session-42.scope: Deactivated successfully. May 15 15:00:04.782200 systemd-logind[1529]: Session 42 logged out. Waiting for processes to exit. May 15 15:00:04.786257 systemd-logind[1529]: Removed session 42. May 15 15:00:07.710400 kubelet[2683]: E0515 15:00:07.710206 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:00:09.784045 systemd[1]: Started sshd@42-24.199.109.202:22-139.178.68.195:49064.service - OpenSSH per-connection server daemon (139.178.68.195:49064). May 15 15:00:09.870430 sshd[5575]: Accepted publickey for core from 139.178.68.195 port 49064 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:09.873153 sshd-session[5575]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:09.880720 systemd-logind[1529]: New session 43 of user core. May 15 15:00:09.885827 systemd[1]: Started session-43.scope - Session 43 of User core. May 15 15:00:10.112772 sshd[5577]: Connection closed by 139.178.68.195 port 49064 May 15 15:00:10.114058 sshd-session[5575]: pam_unix(sshd:session): session closed for user core May 15 15:00:10.120706 systemd[1]: sshd@42-24.199.109.202:22-139.178.68.195:49064.service: Deactivated successfully. May 15 15:00:10.125109 systemd[1]: session-43.scope: Deactivated successfully. May 15 15:00:10.127783 systemd-logind[1529]: Session 43 logged out. Waiting for processes to exit. May 15 15:00:10.130760 systemd-logind[1529]: Removed session 43. May 15 15:00:12.634806 update_engine[1531]: I20250515 15:00:12.634660 1531 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs May 15 15:00:12.634806 update_engine[1531]: I20250515 15:00:12.634766 1531 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs May 15 15:00:12.637247 update_engine[1531]: I20250515 15:00:12.636849 1531 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs May 15 15:00:12.638635 update_engine[1531]: I20250515 15:00:12.638559 1531 omaha_request_params.cc:62] Current group set to developer May 15 15:00:12.638812 update_engine[1531]: I20250515 15:00:12.638772 1531 update_attempter.cc:499] Already updated boot flags. Skipping. May 15 15:00:12.638812 update_engine[1531]: I20250515 15:00:12.638804 1531 update_attempter.cc:643] Scheduling an action processor start. May 15 15:00:12.638909 update_engine[1531]: I20250515 15:00:12.638834 1531 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 15 15:00:12.638970 update_engine[1531]: I20250515 15:00:12.638931 1531 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs May 15 15:00:12.639102 update_engine[1531]: I20250515 15:00:12.639034 1531 omaha_request_action.cc:271] Posting an Omaha request to disabled May 15 15:00:12.639102 update_engine[1531]: I20250515 15:00:12.639060 1531 omaha_request_action.cc:272] Request: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: May 15 15:00:12.639102 update_engine[1531]: I20250515 15:00:12.639070 1531 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 15 15:00:12.657589 update_engine[1531]: I20250515 15:00:12.656065 1531 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 15 15:00:12.657589 update_engine[1531]: I20250515 15:00:12.656688 1531 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 15 15:00:12.659760 update_engine[1531]: E20250515 15:00:12.659502 1531 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 15 15:00:12.659760 update_engine[1531]: I20250515 15:00:12.659708 1531 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 May 15 15:00:12.679718 locksmithd[1572]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 May 15 15:00:15.135894 systemd[1]: Started sshd@43-24.199.109.202:22-139.178.68.195:37278.service - OpenSSH per-connection server daemon (139.178.68.195:37278). May 15 15:00:15.294072 sshd[5589]: Accepted publickey for core from 139.178.68.195 port 37278 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:15.297230 sshd-session[5589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:15.309762 systemd-logind[1529]: New session 44 of user core. May 15 15:00:15.312986 systemd[1]: Started session-44.scope - Session 44 of User core. May 15 15:00:15.531611 sshd[5591]: Connection closed by 139.178.68.195 port 37278 May 15 15:00:15.532366 sshd-session[5589]: pam_unix(sshd:session): session closed for user core May 15 15:00:15.541119 systemd[1]: sshd@43-24.199.109.202:22-139.178.68.195:37278.service: Deactivated successfully. May 15 15:00:15.551604 systemd[1]: session-44.scope: Deactivated successfully. May 15 15:00:15.555730 systemd-logind[1529]: Session 44 logged out. Waiting for processes to exit. May 15 15:00:15.558396 systemd-logind[1529]: Removed session 44. May 15 15:00:18.221459 containerd[1559]: time="2025-05-15T15:00:18.221391589Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"7524dbb4a4abc7e633c5de0b5c3f08ee561a4ef86233337f74cc8e24c9058fff\" pid:5617 exited_at:{seconds:1747321218 nanos:220871554}" May 15 15:00:19.710645 kubelet[2683]: E0515 15:00:19.710374 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:00:20.552906 systemd[1]: Started sshd@44-24.199.109.202:22-139.178.68.195:37284.service - OpenSSH per-connection server daemon (139.178.68.195:37284). May 15 15:00:20.628338 sshd[5630]: Accepted publickey for core from 139.178.68.195 port 37284 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:20.630549 sshd-session[5630]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:20.640248 systemd-logind[1529]: New session 45 of user core. May 15 15:00:20.644857 systemd[1]: Started session-45.scope - Session 45 of User core. May 15 15:00:20.828757 sshd[5632]: Connection closed by 139.178.68.195 port 37284 May 15 15:00:20.830006 sshd-session[5630]: pam_unix(sshd:session): session closed for user core May 15 15:00:20.838586 systemd[1]: sshd@44-24.199.109.202:22-139.178.68.195:37284.service: Deactivated successfully. May 15 15:00:20.842496 systemd[1]: session-45.scope: Deactivated successfully. May 15 15:00:20.845202 systemd-logind[1529]: Session 45 logged out. Waiting for processes to exit. May 15 15:00:20.848228 systemd-logind[1529]: Removed session 45. May 15 15:00:22.519391 update_engine[1531]: I20250515 15:00:22.519274 1531 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 15 15:00:22.519944 update_engine[1531]: I20250515 15:00:22.519665 1531 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 15 15:00:22.520106 update_engine[1531]: I20250515 15:00:22.520061 1531 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 15 15:00:22.523131 update_engine[1531]: E20250515 15:00:22.523053 1531 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 15 15:00:22.523324 update_engine[1531]: I20250515 15:00:22.523200 1531 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 May 15 15:00:25.709975 kubelet[2683]: E0515 15:00:25.709500 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:00:25.855603 systemd[1]: Started sshd@45-24.199.109.202:22-139.178.68.195:37100.service - OpenSSH per-connection server daemon (139.178.68.195:37100). May 15 15:00:25.938574 sshd[5644]: Accepted publickey for core from 139.178.68.195 port 37100 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:25.940655 sshd-session[5644]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:25.949466 systemd-logind[1529]: New session 46 of user core. May 15 15:00:25.957861 systemd[1]: Started session-46.scope - Session 46 of User core. May 15 15:00:26.136591 sshd[5646]: Connection closed by 139.178.68.195 port 37100 May 15 15:00:26.137502 sshd-session[5644]: pam_unix(sshd:session): session closed for user core May 15 15:00:26.146476 systemd[1]: sshd@45-24.199.109.202:22-139.178.68.195:37100.service: Deactivated successfully. May 15 15:00:26.150042 systemd[1]: session-46.scope: Deactivated successfully. May 15 15:00:26.151638 systemd-logind[1529]: Session 46 logged out. Waiting for processes to exit. May 15 15:00:26.155472 systemd-logind[1529]: Removed session 46. May 15 15:00:28.388731 containerd[1559]: time="2025-05-15T15:00:28.388662622Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"65e345359ab6eb35f966ad5bba1eccf882723d04857fb864304fd7a4eefc8c38\" pid:5675 exited_at:{seconds:1747321228 nanos:387919341}" May 15 15:00:28.413695 containerd[1559]: time="2025-05-15T15:00:28.413623141Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"292aeba71cb6563a5bc28a388e3a6a38585c28f81a2819913d5d91f820bdae6b\" pid:5692 exited_at:{seconds:1747321228 nanos:412822693}" May 15 15:00:31.157393 systemd[1]: Started sshd@46-24.199.109.202:22-139.178.68.195:37112.service - OpenSSH per-connection server daemon (139.178.68.195:37112). May 15 15:00:31.240338 sshd[5718]: Accepted publickey for core from 139.178.68.195 port 37112 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:31.244058 sshd-session[5718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:31.254485 systemd-logind[1529]: New session 47 of user core. May 15 15:00:31.258942 systemd[1]: Started session-47.scope - Session 47 of User core. May 15 15:00:31.430653 sshd[5720]: Connection closed by 139.178.68.195 port 37112 May 15 15:00:31.431774 sshd-session[5718]: pam_unix(sshd:session): session closed for user core May 15 15:00:31.440374 systemd-logind[1529]: Session 47 logged out. Waiting for processes to exit. May 15 15:00:31.440919 systemd[1]: sshd@46-24.199.109.202:22-139.178.68.195:37112.service: Deactivated successfully. May 15 15:00:31.444512 systemd[1]: session-47.scope: Deactivated successfully. May 15 15:00:31.448775 systemd-logind[1529]: Removed session 47. May 15 15:00:32.518568 update_engine[1531]: I20250515 15:00:32.518451 1531 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 15 15:00:32.519967 update_engine[1531]: I20250515 15:00:32.519413 1531 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 15 15:00:32.519967 update_engine[1531]: I20250515 15:00:32.519865 1531 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 15 15:00:32.520789 update_engine[1531]: E20250515 15:00:32.520725 1531 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 15 15:00:32.521237 update_engine[1531]: I20250515 15:00:32.521077 1531 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 May 15 15:00:36.470609 systemd[1]: Started sshd@47-24.199.109.202:22-139.178.68.195:59064.service - OpenSSH per-connection server daemon (139.178.68.195:59064). May 15 15:00:36.553288 sshd[5732]: Accepted publickey for core from 139.178.68.195 port 59064 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:36.555877 sshd-session[5732]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:36.563386 systemd-logind[1529]: New session 48 of user core. May 15 15:00:36.573919 systemd[1]: Started session-48.scope - Session 48 of User core. May 15 15:00:36.750235 sshd[5734]: Connection closed by 139.178.68.195 port 59064 May 15 15:00:36.751807 sshd-session[5732]: pam_unix(sshd:session): session closed for user core May 15 15:00:36.759864 systemd[1]: sshd@47-24.199.109.202:22-139.178.68.195:59064.service: Deactivated successfully. May 15 15:00:36.765015 systemd[1]: session-48.scope: Deactivated successfully. May 15 15:00:36.767476 systemd-logind[1529]: Session 48 logged out. Waiting for processes to exit. May 15 15:00:36.770763 systemd-logind[1529]: Removed session 48. May 15 15:00:38.714734 kubelet[2683]: E0515 15:00:38.714611 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:00:39.709573 kubelet[2683]: E0515 15:00:39.709245 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:00:41.769911 systemd[1]: Started sshd@48-24.199.109.202:22-139.178.68.195:59070.service - OpenSSH per-connection server daemon (139.178.68.195:59070). May 15 15:00:41.849630 sshd[5746]: Accepted publickey for core from 139.178.68.195 port 59070 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:41.851652 sshd-session[5746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:41.858294 systemd-logind[1529]: New session 49 of user core. May 15 15:00:41.868815 systemd[1]: Started session-49.scope - Session 49 of User core. May 15 15:00:42.025898 sshd[5748]: Connection closed by 139.178.68.195 port 59070 May 15 15:00:42.026926 sshd-session[5746]: pam_unix(sshd:session): session closed for user core May 15 15:00:42.033188 systemd[1]: sshd@48-24.199.109.202:22-139.178.68.195:59070.service: Deactivated successfully. May 15 15:00:42.036315 systemd[1]: session-49.scope: Deactivated successfully. May 15 15:00:42.037790 systemd-logind[1529]: Session 49 logged out. Waiting for processes to exit. May 15 15:00:42.039312 systemd-logind[1529]: Removed session 49. May 15 15:00:42.517834 update_engine[1531]: I20250515 15:00:42.517714 1531 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 15 15:00:42.518388 update_engine[1531]: I20250515 15:00:42.518011 1531 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 15 15:00:42.518388 update_engine[1531]: I20250515 15:00:42.518360 1531 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 15 15:00:42.519203 update_engine[1531]: E20250515 15:00:42.519162 1531 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 15 15:00:42.519290 update_engine[1531]: I20250515 15:00:42.519213 1531 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 15 15:00:42.519290 update_engine[1531]: I20250515 15:00:42.519222 1531 omaha_request_action.cc:617] Omaha request response: May 15 15:00:42.519348 update_engine[1531]: E20250515 15:00:42.519302 1531 omaha_request_action.cc:636] Omaha request network transfer failed. May 15 15:00:42.519348 update_engine[1531]: I20250515 15:00:42.519330 1531 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. May 15 15:00:42.519348 update_engine[1531]: I20250515 15:00:42.519335 1531 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 15 15:00:42.519348 update_engine[1531]: I20250515 15:00:42.519340 1531 update_attempter.cc:306] Processing Done. May 15 15:00:42.519432 update_engine[1531]: E20250515 15:00:42.519356 1531 update_attempter.cc:619] Update failed. May 15 15:00:42.519432 update_engine[1531]: I20250515 15:00:42.519361 1531 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse May 15 15:00:42.519432 update_engine[1531]: I20250515 15:00:42.519367 1531 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) May 15 15:00:42.519432 update_engine[1531]: I20250515 15:00:42.519373 1531 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. May 15 15:00:42.519845 update_engine[1531]: I20250515 15:00:42.519444 1531 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 15 15:00:42.519845 update_engine[1531]: I20250515 15:00:42.519469 1531 omaha_request_action.cc:271] Posting an Omaha request to disabled May 15 15:00:42.519845 update_engine[1531]: I20250515 15:00:42.519474 1531 omaha_request_action.cc:272] Request: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: May 15 15:00:42.519845 update_engine[1531]: I20250515 15:00:42.519479 1531 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 15 15:00:42.519845 update_engine[1531]: I20250515 15:00:42.519644 1531 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 15 15:00:42.520160 update_engine[1531]: I20250515 15:00:42.519871 1531 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 15 15:00:42.520188 locksmithd[1572]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 May 15 15:00:42.521986 update_engine[1531]: E20250515 15:00:42.521925 1531 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 15 15:00:42.521986 update_engine[1531]: I20250515 15:00:42.521988 1531 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.521996 1531 omaha_request_action.cc:617] Omaha request response: May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.522006 1531 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.522010 1531 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.522015 1531 update_attempter.cc:306] Processing Done. May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.522022 1531 update_attempter.cc:310] Error event sent. May 15 15:00:42.522092 update_engine[1531]: I20250515 15:00:42.522032 1531 update_check_scheduler.cc:74] Next update check in 49m58s May 15 15:00:42.522485 locksmithd[1572]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 May 15 15:00:47.042861 systemd[1]: Started sshd@49-24.199.109.202:22-139.178.68.195:42804.service - OpenSSH per-connection server daemon (139.178.68.195:42804). May 15 15:00:47.122547 sshd[5762]: Accepted publickey for core from 139.178.68.195 port 42804 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:47.125774 sshd-session[5762]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:47.134720 systemd-logind[1529]: New session 50 of user core. May 15 15:00:47.138886 systemd[1]: Started session-50.scope - Session 50 of User core. May 15 15:00:47.317605 sshd[5764]: Connection closed by 139.178.68.195 port 42804 May 15 15:00:47.318673 sshd-session[5762]: pam_unix(sshd:session): session closed for user core May 15 15:00:47.326647 systemd[1]: sshd@49-24.199.109.202:22-139.178.68.195:42804.service: Deactivated successfully. May 15 15:00:47.330335 systemd[1]: session-50.scope: Deactivated successfully. May 15 15:00:47.332648 systemd-logind[1529]: Session 50 logged out. Waiting for processes to exit. May 15 15:00:47.336013 systemd-logind[1529]: Removed session 50. May 15 15:00:48.192081 containerd[1559]: time="2025-05-15T15:00:48.192028362Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"b298874e7c55338a82a0ff0ad067d1be1bafea5253683069166a305cee296204\" pid:5789 exited_at:{seconds:1747321248 nanos:191429258}" May 15 15:00:52.338370 systemd[1]: Started sshd@50-24.199.109.202:22-139.178.68.195:42814.service - OpenSSH per-connection server daemon (139.178.68.195:42814). May 15 15:00:52.526272 sshd[5802]: Accepted publickey for core from 139.178.68.195 port 42814 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:52.529983 sshd-session[5802]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:52.543213 systemd-logind[1529]: New session 51 of user core. May 15 15:00:52.551972 systemd[1]: Started session-51.scope - Session 51 of User core. May 15 15:00:52.983707 sshd[5804]: Connection closed by 139.178.68.195 port 42814 May 15 15:00:52.984560 sshd-session[5802]: pam_unix(sshd:session): session closed for user core May 15 15:00:52.993500 systemd[1]: sshd@50-24.199.109.202:22-139.178.68.195:42814.service: Deactivated successfully. May 15 15:00:52.993618 systemd-logind[1529]: Session 51 logged out. Waiting for processes to exit. May 15 15:00:52.997662 systemd[1]: session-51.scope: Deactivated successfully. May 15 15:00:53.002513 systemd-logind[1529]: Removed session 51. May 15 15:00:58.004617 systemd[1]: Started sshd@51-24.199.109.202:22-139.178.68.195:40164.service - OpenSSH per-connection server daemon (139.178.68.195:40164). May 15 15:00:58.093063 sshd[5816]: Accepted publickey for core from 139.178.68.195 port 40164 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:00:58.095746 sshd-session[5816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:00:58.103389 systemd-logind[1529]: New session 52 of user core. May 15 15:00:58.111942 systemd[1]: Started session-52.scope - Session 52 of User core. May 15 15:00:58.293111 sshd[5818]: Connection closed by 139.178.68.195 port 40164 May 15 15:00:58.294848 sshd-session[5816]: pam_unix(sshd:session): session closed for user core May 15 15:00:58.306761 systemd[1]: sshd@51-24.199.109.202:22-139.178.68.195:40164.service: Deactivated successfully. May 15 15:00:58.310835 systemd[1]: session-52.scope: Deactivated successfully. May 15 15:00:58.314886 systemd-logind[1529]: Session 52 logged out. Waiting for processes to exit. May 15 15:00:58.318075 systemd-logind[1529]: Removed session 52. May 15 15:00:58.406428 containerd[1559]: time="2025-05-15T15:00:58.406284341Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f9c6de2607308ce4d95c7ae8a200e93f8e674cbbb774af3e80973751dc04a50a\" id:\"f0c5bdf9beec90a9645d0f75631671238940b8a98461059744f335e5b81bd4ed\" pid:5840 exited_at:{seconds:1747321258 nanos:405581513}" May 15 15:01:03.326955 systemd[1]: Started sshd@52-24.199.109.202:22-139.178.68.195:40170.service - OpenSSH per-connection server daemon (139.178.68.195:40170). May 15 15:01:03.452742 sshd[5850]: Accepted publickey for core from 139.178.68.195 port 40170 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:01:03.457030 sshd-session[5850]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:01:03.467605 systemd-logind[1529]: New session 53 of user core. May 15 15:01:03.476935 systemd[1]: Started session-53.scope - Session 53 of User core. May 15 15:01:03.677219 sshd[5852]: Connection closed by 139.178.68.195 port 40170 May 15 15:01:03.678876 sshd-session[5850]: pam_unix(sshd:session): session closed for user core May 15 15:01:03.687979 systemd[1]: sshd@52-24.199.109.202:22-139.178.68.195:40170.service: Deactivated successfully. May 15 15:01:03.695038 systemd[1]: session-53.scope: Deactivated successfully. May 15 15:01:03.699881 systemd-logind[1529]: Session 53 logged out. Waiting for processes to exit. May 15 15:01:03.703297 systemd-logind[1529]: Removed session 53. May 15 15:01:04.710548 kubelet[2683]: E0515 15:01:04.709795 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:01:07.709893 kubelet[2683]: E0515 15:01:07.709628 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:01:08.697009 systemd[1]: Started sshd@53-24.199.109.202:22-139.178.68.195:51002.service - OpenSSH per-connection server daemon (139.178.68.195:51002). May 15 15:01:08.783060 sshd[5864]: Accepted publickey for core from 139.178.68.195 port 51002 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:01:08.786940 sshd-session[5864]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:01:08.794578 systemd-logind[1529]: New session 54 of user core. May 15 15:01:08.805485 systemd[1]: Started session-54.scope - Session 54 of User core. May 15 15:01:09.019801 sshd[5866]: Connection closed by 139.178.68.195 port 51002 May 15 15:01:09.022304 sshd-session[5864]: pam_unix(sshd:session): session closed for user core May 15 15:01:09.029486 systemd[1]: sshd@53-24.199.109.202:22-139.178.68.195:51002.service: Deactivated successfully. May 15 15:01:09.032719 systemd[1]: session-54.scope: Deactivated successfully. May 15 15:01:09.034873 systemd-logind[1529]: Session 54 logged out. Waiting for processes to exit. May 15 15:01:09.038024 systemd-logind[1529]: Removed session 54. May 15 15:01:14.053058 systemd[1]: Started sshd@54-24.199.109.202:22-139.178.68.195:58760.service - OpenSSH per-connection server daemon (139.178.68.195:58760). May 15 15:01:14.172347 sshd[5879]: Accepted publickey for core from 139.178.68.195 port 58760 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:01:14.174749 sshd-session[5879]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:01:14.185715 systemd-logind[1529]: New session 55 of user core. May 15 15:01:14.192042 systemd[1]: Started session-55.scope - Session 55 of User core. May 15 15:01:14.424089 sshd[5881]: Connection closed by 139.178.68.195 port 58760 May 15 15:01:14.423847 sshd-session[5879]: pam_unix(sshd:session): session closed for user core May 15 15:01:14.433674 systemd[1]: sshd@54-24.199.109.202:22-139.178.68.195:58760.service: Deactivated successfully. May 15 15:01:14.437688 systemd[1]: session-55.scope: Deactivated successfully. May 15 15:01:14.438928 systemd-logind[1529]: Session 55 logged out. Waiting for processes to exit. May 15 15:01:14.441148 systemd-logind[1529]: Removed session 55. May 15 15:01:15.710562 kubelet[2683]: E0515 15:01:15.710325 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:01:15.710562 kubelet[2683]: E0515 15:01:15.710325 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 15 15:01:18.182183 containerd[1559]: time="2025-05-15T15:01:18.181923708Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8118f5ebd528abfdde243c0cb7d1208f786f9e53163701270bf2fa0243289b5f\" id:\"b62ae2bab786a7aa02bbfbc39020ee9f229857ec2436f3838417ed642f52b427\" pid:5907 exited_at:{seconds:1747321278 nanos:181466679}" May 15 15:01:19.439425 systemd[1]: Started sshd@55-24.199.109.202:22-139.178.68.195:58768.service - OpenSSH per-connection server daemon (139.178.68.195:58768). May 15 15:01:19.530053 sshd[5920]: Accepted publickey for core from 139.178.68.195 port 58768 ssh2: RSA SHA256:CR2QFGI8Wi38j7m0fVendNlhmaPvJh+gYMXcH5yQYrY May 15 15:01:19.533015 sshd-session[5920]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 15 15:01:19.543139 systemd-logind[1529]: New session 56 of user core. May 15 15:01:19.545824 systemd[1]: Started session-56.scope - Session 56 of User core. May 15 15:01:19.846903 sshd[5922]: Connection closed by 139.178.68.195 port 58768 May 15 15:01:19.847716 sshd-session[5920]: pam_unix(sshd:session): session closed for user core May 15 15:01:19.855313 systemd-logind[1529]: Session 56 logged out. Waiting for processes to exit. May 15 15:01:19.856808 systemd[1]: sshd@55-24.199.109.202:22-139.178.68.195:58768.service: Deactivated successfully. May 15 15:01:19.859261 systemd[1]: session-56.scope: Deactivated successfully. May 15 15:01:19.861769 systemd-logind[1529]: Removed session 56.