Sep 4 17:40:35.094196 kernel: Linux version 6.6.48-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Wed Sep 4 15:54:07 -00 2024 Sep 4 17:40:35.094225 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.094236 kernel: BIOS-provided physical RAM map: Sep 4 17:40:35.094242 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 4 17:40:35.094249 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Sep 4 17:40:35.094257 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ff40fff] usable Sep 4 17:40:35.094264 kernel: BIOS-e820: [mem 0x000000003ff41000-0x000000003ff70fff] type 20 Sep 4 17:40:35.094276 kernel: BIOS-e820: [mem 0x000000003ff71000-0x000000003ffc8fff] reserved Sep 4 17:40:35.094282 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Sep 4 17:40:35.094289 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Sep 4 17:40:35.094298 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Sep 4 17:40:35.094304 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Sep 4 17:40:35.094311 kernel: printk: bootconsole [earlyser0] enabled Sep 4 17:40:35.094320 kernel: NX (Execute Disable) protection: active Sep 4 17:40:35.094330 kernel: APIC: Static calls initialized Sep 4 17:40:35.094340 kernel: efi: EFI v2.7 by Microsoft Sep 4 17:40:35.094348 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff85000 SMBIOS 3.0=0x3ff83000 MEMATTR=0x3f5c1a98 Sep 4 17:40:35.094355 kernel: SMBIOS 3.1.0 present. Sep 4 17:40:35.094365 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 03/08/2024 Sep 4 17:40:35.094372 kernel: Hypervisor detected: Microsoft Hyper-V Sep 4 17:40:35.094380 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0x64e24, misc 0xbed7b2 Sep 4 17:40:35.094389 kernel: Hyper-V: Host Build 10.0.20348.1633-1-0 Sep 4 17:40:35.094396 kernel: Hyper-V: Nested features: 0x1e0101 Sep 4 17:40:35.094406 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Sep 4 17:40:35.094416 kernel: Hyper-V: Using hypercall for remote TLB flush Sep 4 17:40:35.094426 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Sep 4 17:40:35.094434 kernel: clocksource: hyperv_clocksource_msr: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Sep 4 17:40:35.094442 kernel: tsc: Marking TSC unstable due to running on Hyper-V Sep 4 17:40:35.094452 kernel: tsc: Detected 2593.907 MHz processor Sep 4 17:40:35.094459 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 4 17:40:35.094469 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 4 17:40:35.094478 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x400000000 Sep 4 17:40:35.094486 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 4 17:40:35.094497 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 4 17:40:35.094504 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved Sep 4 17:40:35.094513 kernel: last_pfn = 0x40000 max_arch_pfn = 0x400000000 Sep 4 17:40:35.094521 kernel: Using GB pages for direct mapping Sep 4 17:40:35.094528 kernel: Secure boot disabled Sep 4 17:40:35.094544 kernel: ACPI: Early table checksum verification disabled Sep 4 17:40:35.094554 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Sep 4 17:40:35.094569 kernel: ACPI: XSDT 0x000000003FFF90E8 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094581 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094590 kernel: ACPI: DSDT 0x000000003FFD6000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Sep 4 17:40:35.094599 kernel: ACPI: FACS 0x000000003FFFE000 000040 Sep 4 17:40:35.094610 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094618 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094629 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094639 kernel: ACPI: APIC 0x000000003FFD5000 000058 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094649 kernel: ACPI: SRAT 0x000000003FFD4000 0002D0 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094657 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094665 kernel: ACPI: FPDT 0x000000003FFD2000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094676 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Sep 4 17:40:35.094683 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4183] Sep 4 17:40:35.094694 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Sep 4 17:40:35.094702 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Sep 4 17:40:35.094714 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Sep 4 17:40:35.094722 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Sep 4 17:40:35.094730 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5057] Sep 4 17:40:35.094741 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd42cf] Sep 4 17:40:35.094748 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Sep 4 17:40:35.094759 kernel: ACPI: Reserving FPDT table memory at [mem 0x3ffd2000-0x3ffd2033] Sep 4 17:40:35.094767 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 4 17:40:35.094775 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 4 17:40:35.094784 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Sep 4 17:40:35.094795 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] hotplug Sep 4 17:40:35.094805 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2c0000000-0xfdfffffff] hotplug Sep 4 17:40:35.094812 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Sep 4 17:40:35.094823 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Sep 4 17:40:35.094831 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Sep 4 17:40:35.094840 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Sep 4 17:40:35.094849 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Sep 4 17:40:35.094856 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Sep 4 17:40:35.094867 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Sep 4 17:40:35.094877 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Sep 4 17:40:35.094888 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Sep 4 17:40:35.094896 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000000-0x1ffffffffffff] hotplug Sep 4 17:40:35.094904 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2000000000000-0x3ffffffffffff] hotplug Sep 4 17:40:35.094914 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x4000000000000-0x7ffffffffffff] hotplug Sep 4 17:40:35.094921 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x8000000000000-0xfffffffffffff] hotplug Sep 4 17:40:35.094937 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00000000-0x2bfffffff] Sep 4 17:40:35.094948 kernel: NODE_DATA(0) allocated [mem 0x2bfffa000-0x2bfffffff] Sep 4 17:40:35.094956 kernel: Zone ranges: Sep 4 17:40:35.094969 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 4 17:40:35.094979 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Sep 4 17:40:35.094989 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Sep 4 17:40:35.094999 kernel: Movable zone start for each node Sep 4 17:40:35.095007 kernel: Early memory node ranges Sep 4 17:40:35.095015 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 4 17:40:35.095025 kernel: node 0: [mem 0x0000000000100000-0x000000003ff40fff] Sep 4 17:40:35.095033 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Sep 4 17:40:35.095043 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Sep 4 17:40:35.095053 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Sep 4 17:40:35.095064 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 4 17:40:35.095072 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 4 17:40:35.095089 kernel: On node 0, zone DMA32: 190 pages in unavailable ranges Sep 4 17:40:35.095097 kernel: ACPI: PM-Timer IO Port: 0x408 Sep 4 17:40:35.095108 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) Sep 4 17:40:35.095115 kernel: IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 Sep 4 17:40:35.095124 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 4 17:40:35.095134 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 4 17:40:35.095145 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Sep 4 17:40:35.095154 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 4 17:40:35.095162 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Sep 4 17:40:35.095172 kernel: Booting paravirtualized kernel on Hyper-V Sep 4 17:40:35.095180 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 4 17:40:35.095190 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 4 17:40:35.095199 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u1048576 Sep 4 17:40:35.095206 kernel: pcpu-alloc: s196904 r8192 d32472 u1048576 alloc=1*2097152 Sep 4 17:40:35.095217 kernel: pcpu-alloc: [0] 0 1 Sep 4 17:40:35.095226 kernel: Hyper-V: PV spinlocks enabled Sep 4 17:40:35.095237 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 4 17:40:35.095246 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.095257 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 4 17:40:35.095265 kernel: random: crng init done Sep 4 17:40:35.095273 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Sep 4 17:40:35.095283 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 4 17:40:35.095290 kernel: Fallback order for Node 0: 0 Sep 4 17:40:35.095303 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2062618 Sep 4 17:40:35.095326 kernel: Policy zone: Normal Sep 4 17:40:35.095340 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 4 17:40:35.095350 kernel: software IO TLB: area num 2. Sep 4 17:40:35.095360 kernel: Memory: 8077076K/8387460K available (12288K kernel code, 2304K rwdata, 22708K rodata, 42704K init, 2488K bss, 310124K reserved, 0K cma-reserved) Sep 4 17:40:35.095371 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 4 17:40:35.095381 kernel: ftrace: allocating 37748 entries in 148 pages Sep 4 17:40:35.095392 kernel: ftrace: allocated 148 pages with 3 groups Sep 4 17:40:35.095400 kernel: Dynamic Preempt: voluntary Sep 4 17:40:35.095410 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 4 17:40:35.095420 kernel: rcu: RCU event tracing is enabled. Sep 4 17:40:35.095433 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 4 17:40:35.095443 kernel: Trampoline variant of Tasks RCU enabled. Sep 4 17:40:35.095452 kernel: Rude variant of Tasks RCU enabled. Sep 4 17:40:35.095462 kernel: Tracing variant of Tasks RCU enabled. Sep 4 17:40:35.095470 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 4 17:40:35.095483 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 4 17:40:35.095492 kernel: Using NULL legacy PIC Sep 4 17:40:35.095503 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Sep 4 17:40:35.095511 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 4 17:40:35.095522 kernel: Console: colour dummy device 80x25 Sep 4 17:40:35.095530 kernel: printk: console [tty1] enabled Sep 4 17:40:35.095540 kernel: printk: console [ttyS0] enabled Sep 4 17:40:35.095549 kernel: printk: bootconsole [earlyser0] disabled Sep 4 17:40:35.095558 kernel: ACPI: Core revision 20230628 Sep 4 17:40:35.095569 kernel: Failed to register legacy timer interrupt Sep 4 17:40:35.095579 kernel: APIC: Switch to symmetric I/O mode setup Sep 4 17:40:35.095590 kernel: Hyper-V: enabling crash_kexec_post_notifiers Sep 4 17:40:35.095598 kernel: Hyper-V: Using IPI hypercalls Sep 4 17:40:35.099814 kernel: APIC: send_IPI() replaced with hv_send_ipi() Sep 4 17:40:35.099840 kernel: APIC: send_IPI_mask() replaced with hv_send_ipi_mask() Sep 4 17:40:35.099854 kernel: APIC: send_IPI_mask_allbutself() replaced with hv_send_ipi_mask_allbutself() Sep 4 17:40:35.099869 kernel: APIC: send_IPI_allbutself() replaced with hv_send_ipi_allbutself() Sep 4 17:40:35.099884 kernel: APIC: send_IPI_all() replaced with hv_send_ipi_all() Sep 4 17:40:35.099898 kernel: APIC: send_IPI_self() replaced with hv_send_ipi_self() Sep 4 17:40:35.099917 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 5187.81 BogoMIPS (lpj=2593907) Sep 4 17:40:35.099932 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 4 17:40:35.099946 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Sep 4 17:40:35.099960 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 4 17:40:35.099973 kernel: Spectre V2 : Mitigation: Retpolines Sep 4 17:40:35.099986 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Sep 4 17:40:35.100000 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Sep 4 17:40:35.100012 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Sep 4 17:40:35.100029 kernel: RETBleed: Vulnerable Sep 4 17:40:35.100053 kernel: Speculative Store Bypass: Vulnerable Sep 4 17:40:35.100066 kernel: TAA: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:40:35.100087 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:40:35.100101 kernel: GDS: Unknown: Dependent on hypervisor status Sep 4 17:40:35.100115 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 4 17:40:35.100127 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 4 17:40:35.100140 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 4 17:40:35.100154 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Sep 4 17:40:35.100168 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Sep 4 17:40:35.100182 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Sep 4 17:40:35.100196 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 4 17:40:35.100214 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Sep 4 17:40:35.100229 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Sep 4 17:40:35.100244 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Sep 4 17:40:35.100260 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Sep 4 17:40:35.100275 kernel: Freeing SMP alternatives memory: 32K Sep 4 17:40:35.100290 kernel: pid_max: default: 32768 minimum: 301 Sep 4 17:40:35.100304 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 4 17:40:35.100319 kernel: landlock: Up and running. Sep 4 17:40:35.100334 kernel: SELinux: Initializing. Sep 4 17:40:35.100349 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.100363 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.100378 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x7) Sep 4 17:40:35.100396 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100411 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100427 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100442 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Sep 4 17:40:35.100457 kernel: signal: max sigframe size: 3632 Sep 4 17:40:35.100472 kernel: rcu: Hierarchical SRCU implementation. Sep 4 17:40:35.100487 kernel: rcu: Max phase no-delay instances is 400. Sep 4 17:40:35.100503 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 4 17:40:35.100518 kernel: smp: Bringing up secondary CPUs ... Sep 4 17:40:35.100536 kernel: smpboot: x86: Booting SMP configuration: Sep 4 17:40:35.100551 kernel: .... node #0, CPUs: #1 Sep 4 17:40:35.100564 kernel: TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. Sep 4 17:40:35.100578 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 4 17:40:35.100593 kernel: smp: Brought up 1 node, 2 CPUs Sep 4 17:40:35.100607 kernel: smpboot: Max logical packages: 1 Sep 4 17:40:35.100622 kernel: smpboot: Total of 2 processors activated (10375.62 BogoMIPS) Sep 4 17:40:35.100637 kernel: devtmpfs: initialized Sep 4 17:40:35.100654 kernel: x86/mm: Memory block size: 128MB Sep 4 17:40:35.100669 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Sep 4 17:40:35.100684 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 4 17:40:35.100699 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 4 17:40:35.100713 kernel: pinctrl core: initialized pinctrl subsystem Sep 4 17:40:35.100727 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 4 17:40:35.100741 kernel: audit: initializing netlink subsys (disabled) Sep 4 17:40:35.100755 kernel: audit: type=2000 audit(1725471634.028:1): state=initialized audit_enabled=0 res=1 Sep 4 17:40:35.100769 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 4 17:40:35.100786 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 4 17:40:35.100800 kernel: cpuidle: using governor menu Sep 4 17:40:35.100814 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 4 17:40:35.100828 kernel: dca service started, version 1.12.1 Sep 4 17:40:35.100842 kernel: e820: reserve RAM buffer [mem 0x3ff41000-0x3fffffff] Sep 4 17:40:35.100857 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 4 17:40:35.100871 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 4 17:40:35.100885 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 4 17:40:35.100899 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 4 17:40:35.100915 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 4 17:40:35.100930 kernel: ACPI: Added _OSI(Module Device) Sep 4 17:40:35.100944 kernel: ACPI: Added _OSI(Processor Device) Sep 4 17:40:35.100958 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Sep 4 17:40:35.100972 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 4 17:40:35.100986 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 4 17:40:35.101000 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 4 17:40:35.101014 kernel: ACPI: Interpreter enabled Sep 4 17:40:35.101028 kernel: ACPI: PM: (supports S0 S5) Sep 4 17:40:35.101045 kernel: ACPI: Using IOAPIC for interrupt routing Sep 4 17:40:35.101059 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 4 17:40:35.101073 kernel: PCI: Ignoring E820 reservations for host bridge windows Sep 4 17:40:35.101104 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Sep 4 17:40:35.101118 kernel: iommu: Default domain type: Translated Sep 4 17:40:35.101133 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 4 17:40:35.101147 kernel: efivars: Registered efivars operations Sep 4 17:40:35.101160 kernel: PCI: Using ACPI for IRQ routing Sep 4 17:40:35.101175 kernel: PCI: System does not support PCI Sep 4 17:40:35.101191 kernel: vgaarb: loaded Sep 4 17:40:35.101205 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page Sep 4 17:40:35.101219 kernel: VFS: Disk quotas dquot_6.6.0 Sep 4 17:40:35.101233 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 4 17:40:35.101248 kernel: pnp: PnP ACPI init Sep 4 17:40:35.101261 kernel: pnp: PnP ACPI: found 3 devices Sep 4 17:40:35.101275 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 4 17:40:35.101290 kernel: NET: Registered PF_INET protocol family Sep 4 17:40:35.101304 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 4 17:40:35.101320 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Sep 4 17:40:35.101335 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 4 17:40:35.101348 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 4 17:40:35.101363 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Sep 4 17:40:35.101377 kernel: TCP: Hash tables configured (established 65536 bind 65536) Sep 4 17:40:35.101391 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.101405 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.101419 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 4 17:40:35.101433 kernel: NET: Registered PF_XDP protocol family Sep 4 17:40:35.101449 kernel: PCI: CLS 0 bytes, default 64 Sep 4 17:40:35.101463 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 4 17:40:35.101476 kernel: software IO TLB: mapped [mem 0x000000003b5c1000-0x000000003f5c1000] (64MB) Sep 4 17:40:35.101488 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 4 17:40:35.101503 kernel: Initialise system trusted keyrings Sep 4 17:40:35.101516 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Sep 4 17:40:35.101530 kernel: Key type asymmetric registered Sep 4 17:40:35.101543 kernel: Asymmetric key parser 'x509' registered Sep 4 17:40:35.101572 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 4 17:40:35.101589 kernel: io scheduler mq-deadline registered Sep 4 17:40:35.101601 kernel: io scheduler kyber registered Sep 4 17:40:35.101615 kernel: io scheduler bfq registered Sep 4 17:40:35.101630 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 4 17:40:35.101647 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 4 17:40:35.101661 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 4 17:40:35.101675 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Sep 4 17:40:35.101689 kernel: i8042: PNP: No PS/2 controller found. Sep 4 17:40:35.101865 kernel: rtc_cmos 00:02: registered as rtc0 Sep 4 17:40:35.101994 kernel: rtc_cmos 00:02: setting system clock to 2024-09-04T17:40:34 UTC (1725471634) Sep 4 17:40:35.102136 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Sep 4 17:40:35.102153 kernel: intel_pstate: CPU model not supported Sep 4 17:40:35.102167 kernel: efifb: probing for efifb Sep 4 17:40:35.102181 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Sep 4 17:40:35.102195 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Sep 4 17:40:35.102209 kernel: efifb: scrolling: redraw Sep 4 17:40:35.102223 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 4 17:40:35.102240 kernel: Console: switching to colour frame buffer device 128x48 Sep 4 17:40:35.102254 kernel: fb0: EFI VGA frame buffer device Sep 4 17:40:35.102268 kernel: pstore: Using crash dump compression: deflate Sep 4 17:40:35.102282 kernel: pstore: Registered efi_pstore as persistent store backend Sep 4 17:40:35.102296 kernel: NET: Registered PF_INET6 protocol family Sep 4 17:40:35.102309 kernel: Segment Routing with IPv6 Sep 4 17:40:35.102323 kernel: In-situ OAM (IOAM) with IPv6 Sep 4 17:40:35.102336 kernel: NET: Registered PF_PACKET protocol family Sep 4 17:40:35.102350 kernel: Key type dns_resolver registered Sep 4 17:40:35.102366 kernel: IPI shorthand broadcast: enabled Sep 4 17:40:35.102390 kernel: sched_clock: Marking stable (851002800, 50107800)->(1128054500, -226943900) Sep 4 17:40:35.102403 kernel: registered taskstats version 1 Sep 4 17:40:35.102417 kernel: Loading compiled-in X.509 certificates Sep 4 17:40:35.102429 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.48-flatcar: 8669771ab5e11f458b79e6634fe685dacc266b18' Sep 4 17:40:35.102442 kernel: Key type .fscrypt registered Sep 4 17:40:35.102455 kernel: Key type fscrypt-provisioning registered Sep 4 17:40:35.102467 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 4 17:40:35.102483 kernel: ima: Allocated hash algorithm: sha1 Sep 4 17:40:35.102496 kernel: ima: No architecture policies found Sep 4 17:40:35.102509 kernel: clk: Disabling unused clocks Sep 4 17:40:35.102522 kernel: Freeing unused kernel image (initmem) memory: 42704K Sep 4 17:40:35.102535 kernel: Write protecting the kernel read-only data: 36864k Sep 4 17:40:35.102548 kernel: Freeing unused kernel image (rodata/data gap) memory: 1868K Sep 4 17:40:35.102561 kernel: Run /init as init process Sep 4 17:40:35.102574 kernel: with arguments: Sep 4 17:40:35.102587 kernel: /init Sep 4 17:40:35.102599 kernel: with environment: Sep 4 17:40:35.102614 kernel: HOME=/ Sep 4 17:40:35.102630 kernel: TERM=linux Sep 4 17:40:35.102643 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 4 17:40:35.102657 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:40:35.102673 systemd[1]: Detected virtualization microsoft. Sep 4 17:40:35.102686 systemd[1]: Detected architecture x86-64. Sep 4 17:40:35.102699 systemd[1]: Running in initrd. Sep 4 17:40:35.102722 systemd[1]: No hostname configured, using default hostname. Sep 4 17:40:35.102735 systemd[1]: Hostname set to . Sep 4 17:40:35.102749 systemd[1]: Initializing machine ID from random generator. Sep 4 17:40:35.102763 systemd[1]: Queued start job for default target initrd.target. Sep 4 17:40:35.102780 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:40:35.102794 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:40:35.102808 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 4 17:40:35.102824 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:40:35.102842 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 4 17:40:35.102857 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 4 17:40:35.102874 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 4 17:40:35.102889 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 4 17:40:35.102904 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:40:35.102918 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:40:35.102933 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:40:35.102950 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:40:35.102965 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:40:35.102979 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:40:35.102994 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:40:35.103008 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:40:35.103022 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 17:40:35.103037 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 4 17:40:35.103052 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:40:35.103066 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:40:35.103699 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:40:35.103718 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:40:35.103734 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 4 17:40:35.103751 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:40:35.103770 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 4 17:40:35.103785 systemd[1]: Starting systemd-fsck-usr.service... Sep 4 17:40:35.103799 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:40:35.103814 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:40:35.103865 systemd-journald[176]: Collecting audit messages is disabled. Sep 4 17:40:35.103898 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:35.103913 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 4 17:40:35.103945 systemd-journald[176]: Journal started Sep 4 17:40:35.103985 systemd-journald[176]: Runtime Journal (/run/log/journal/ee2e414552134498a0bdd253f0c95dcc) is 8.0M, max 158.8M, 150.8M free. Sep 4 17:40:35.100067 systemd-modules-load[177]: Inserted module 'overlay' Sep 4 17:40:35.115934 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:40:35.119340 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:40:35.122786 systemd[1]: Finished systemd-fsck-usr.service. Sep 4 17:40:35.131778 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:35.144106 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 4 17:40:35.146260 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:35.152068 kernel: Bridge firewalling registered Sep 4 17:40:35.148597 systemd-modules-load[177]: Inserted module 'br_netfilter' Sep 4 17:40:35.162246 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:40:35.171822 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 17:40:35.175190 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:40:35.184973 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:35.188855 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:40:35.192516 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:40:35.212228 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 4 17:40:35.222269 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:40:35.228068 dracut-cmdline[204]: dracut-dracut-053 Sep 4 17:40:35.229951 dracut-cmdline[204]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.249374 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:40:35.262691 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:40:35.277521 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:40:35.283639 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:40:35.330097 kernel: SCSI subsystem initialized Sep 4 17:40:35.330619 systemd-resolved[260]: Positive Trust Anchors: Sep 4 17:40:35.330635 systemd-resolved[260]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:40:35.330683 systemd-resolved[260]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 17:40:35.357050 systemd-resolved[260]: Defaulting to hostname 'linux'. Sep 4 17:40:35.365241 kernel: Loading iSCSI transport class v2.0-870. Sep 4 17:40:35.358253 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:40:35.368057 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:40:35.387101 kernel: iscsi: registered transport (tcp) Sep 4 17:40:35.408086 kernel: iscsi: registered transport (qla4xxx) Sep 4 17:40:35.408154 kernel: QLogic iSCSI HBA Driver Sep 4 17:40:35.444621 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 4 17:40:35.456333 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 4 17:40:35.486719 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 4 17:40:35.486790 kernel: device-mapper: uevent: version 1.0.3 Sep 4 17:40:35.490054 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 4 17:40:35.530101 kernel: raid6: avx512x4 gen() 18093 MB/s Sep 4 17:40:35.549090 kernel: raid6: avx512x2 gen() 18049 MB/s Sep 4 17:40:35.568087 kernel: raid6: avx512x1 gen() 17920 MB/s Sep 4 17:40:35.587092 kernel: raid6: avx2x4 gen() 18120 MB/s Sep 4 17:40:35.605088 kernel: raid6: avx2x2 gen() 17980 MB/s Sep 4 17:40:35.625548 kernel: raid6: avx2x1 gen() 13880 MB/s Sep 4 17:40:35.625607 kernel: raid6: using algorithm avx2x4 gen() 18120 MB/s Sep 4 17:40:35.648197 kernel: raid6: .... xor() 6986 MB/s, rmw enabled Sep 4 17:40:35.648234 kernel: raid6: using avx512x2 recovery algorithm Sep 4 17:40:35.670102 kernel: xor: automatically using best checksumming function avx Sep 4 17:40:35.822104 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 4 17:40:35.831946 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:40:35.841233 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:40:35.853435 systemd-udevd[396]: Using default interface naming scheme 'v255'. Sep 4 17:40:35.857899 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:40:35.873206 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 4 17:40:35.884630 dracut-pre-trigger[406]: rd.md=0: removing MD RAID activation Sep 4 17:40:35.909714 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:40:35.920197 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:40:35.960636 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:40:35.974299 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 4 17:40:36.000406 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 4 17:40:36.010595 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:40:36.010680 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:40:36.011591 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:40:36.034471 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 4 17:40:36.058718 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:40:36.071519 kernel: cryptd: max_cpu_qlen set to 1000 Sep 4 17:40:36.082851 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:40:36.086201 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:36.096120 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:36.108370 kernel: AVX2 version of gcm_enc/dec engaged. Sep 4 17:40:36.108423 kernel: AES CTR mode by8 optimization enabled Sep 4 17:40:36.105116 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:36.118219 kernel: hv_vmbus: Vmbus version:5.2 Sep 4 17:40:36.105377 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.108842 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.121558 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.129055 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:36.129155 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.150168 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 4 17:40:36.150213 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 4 17:40:36.154119 kernel: hv_vmbus: registering driver hyperv_keyboard Sep 4 17:40:36.155428 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.177264 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.185259 kernel: PTP clock support registered Sep 4 17:40:36.189264 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:37.418381 kernel: hv_utils: Registering HyperV Utility Driver Sep 4 17:40:37.418400 kernel: hv_vmbus: registering driver hv_utils Sep 4 17:40:37.418414 kernel: hv_utils: Heartbeat IC version 3.0 Sep 4 17:40:37.418425 kernel: hv_utils: Shutdown IC version 3.2 Sep 4 17:40:37.418438 kernel: hv_utils: TimeSync IC version 4.0 Sep 4 17:40:37.418449 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Sep 4 17:40:37.418460 kernel: hv_vmbus: registering driver hv_storvsc Sep 4 17:40:37.418479 kernel: scsi host1: storvsc_host_t Sep 4 17:40:37.418630 kernel: scsi host0: storvsc_host_t Sep 4 17:40:37.412166 systemd-resolved[260]: Clock change detected. Flushing caches. Sep 4 17:40:37.438117 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Sep 4 17:40:37.445455 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Sep 4 17:40:37.445498 kernel: hv_vmbus: registering driver hv_netvsc Sep 4 17:40:37.449099 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 4 17:40:37.461934 kernel: hv_vmbus: registering driver hid_hyperv Sep 4 17:40:37.469391 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Sep 4 17:40:37.469429 kernel: hid 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Sep 4 17:40:37.476703 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:37.497092 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Sep 4 17:40:37.497317 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 4 17:40:37.503903 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Sep 4 17:40:37.517101 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Sep 4 17:40:37.517393 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Sep 4 17:40:37.517909 kernel: sd 0:0:0:0: [sda] Write Protect is off Sep 4 17:40:37.523133 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Sep 4 17:40:37.523398 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Sep 4 17:40:37.529904 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:37.529939 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Sep 4 17:40:37.655999 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: VF slot 1 added Sep 4 17:40:37.663929 kernel: hv_vmbus: registering driver hv_pci Sep 4 17:40:37.667922 kernel: hv_pci 879eec16-c78a-4323-af28-425ed18503dc: PCI VMBus probing: Using version 0x10004 Sep 4 17:40:37.672919 kernel: hv_pci 879eec16-c78a-4323-af28-425ed18503dc: PCI host bridge to bus c78a:00 Sep 4 17:40:37.673079 kernel: pci_bus c78a:00: root bus resource [mem 0xfe0000000-0xfe00fffff window] Sep 4 17:40:37.678078 kernel: pci_bus c78a:00: No busn resource found for root bus, will use [bus 00-ff] Sep 4 17:40:37.683212 kernel: pci c78a:00:02.0: [15b3:1016] type 00 class 0x020000 Sep 4 17:40:37.686897 kernel: pci c78a:00:02.0: reg 0x10: [mem 0xfe0000000-0xfe00fffff 64bit pref] Sep 4 17:40:37.690947 kernel: pci c78a:00:02.0: enabling Extended Tags Sep 4 17:40:37.701920 kernel: pci c78a:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at c78a:00:02.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Sep 4 17:40:37.708396 kernel: pci_bus c78a:00: busn_res: [bus 00-ff] end is updated to 00 Sep 4 17:40:37.708648 kernel: pci c78a:00:02.0: BAR 0: assigned [mem 0xfe0000000-0xfe00fffff 64bit pref] Sep 4 17:40:37.883429 kernel: mlx5_core c78a:00:02.0: enabling device (0000 -> 0002) Sep 4 17:40:37.887912 kernel: mlx5_core c78a:00:02.0: firmware version: 14.30.1284 Sep 4 17:40:38.034682 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Sep 4 17:40:38.083936 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (455) Sep 4 17:40:38.100349 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Sep 4 17:40:38.121101 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: VF registering: eth1 Sep 4 17:40:38.121342 kernel: mlx5_core c78a:00:02.0 eth1: joined to eth0 Sep 4 17:40:38.125722 kernel: mlx5_core c78a:00:02.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 4 17:40:38.135905 kernel: mlx5_core c78a:00:02.0 enP51082s1: renamed from eth1 Sep 4 17:40:38.147918 kernel: BTRFS: device fsid 0dc40443-7f77-4fa7-b5e4-579d4bba0772 devid 1 transid 37 /dev/sda3 scanned by (udev-worker) (450) Sep 4 17:40:38.172219 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Sep 4 17:40:38.180533 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Sep 4 17:40:38.189459 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Sep 4 17:40:38.199037 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 4 17:40:38.213903 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:38.223906 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:39.230967 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:39.231625 disk-uuid[597]: The operation has completed successfully. Sep 4 17:40:39.313275 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 4 17:40:39.313386 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 4 17:40:39.334063 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 4 17:40:39.340268 sh[683]: Success Sep 4 17:40:39.373155 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 4 17:40:39.595706 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 4 17:40:39.610015 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 4 17:40:39.615597 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 4 17:40:39.633902 kernel: BTRFS info (device dm-0): first mount of filesystem 0dc40443-7f77-4fa7-b5e4-579d4bba0772 Sep 4 17:40:39.633941 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:39.639803 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 4 17:40:39.642882 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 4 17:40:39.645405 kernel: BTRFS info (device dm-0): using free space tree Sep 4 17:40:39.928280 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 4 17:40:39.935986 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 4 17:40:39.945044 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 4 17:40:39.953064 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 4 17:40:39.977560 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:39.977612 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:39.977632 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:39.996906 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:40.012166 kernel: BTRFS info (device sda6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:40.011760 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 4 17:40:40.019446 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 4 17:40:40.031096 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 4 17:40:40.053374 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:40:40.066046 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:40:40.085631 systemd-networkd[867]: lo: Link UP Sep 4 17:40:40.085641 systemd-networkd[867]: lo: Gained carrier Sep 4 17:40:40.087706 systemd-networkd[867]: Enumeration completed Sep 4 17:40:40.088298 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:40:40.092252 systemd-networkd[867]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:40.092256 systemd-networkd[867]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:40:40.093017 systemd[1]: Reached target network.target - Network. Sep 4 17:40:40.159952 kernel: mlx5_core c78a:00:02.0 enP51082s1: Link up Sep 4 17:40:40.194923 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: Data path switched to VF: enP51082s1 Sep 4 17:40:40.195994 systemd-networkd[867]: enP51082s1: Link UP Sep 4 17:40:40.196160 systemd-networkd[867]: eth0: Link UP Sep 4 17:40:40.196406 systemd-networkd[867]: eth0: Gained carrier Sep 4 17:40:40.196420 systemd-networkd[867]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:40.211150 systemd-networkd[867]: enP51082s1: Gained carrier Sep 4 17:40:40.232993 systemd-networkd[867]: eth0: DHCPv4 address 10.200.4.34/24, gateway 10.200.4.1 acquired from 168.63.129.16 Sep 4 17:40:40.963967 ignition[846]: Ignition 2.19.0 Sep 4 17:40:40.963979 ignition[846]: Stage: fetch-offline Sep 4 17:40:40.964024 ignition[846]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:40.964036 ignition[846]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:40.964147 ignition[846]: parsed url from cmdline: "" Sep 4 17:40:40.964151 ignition[846]: no config URL provided Sep 4 17:40:40.964159 ignition[846]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:40:40.964170 ignition[846]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:40:40.964176 ignition[846]: failed to fetch config: resource requires networking Sep 4 17:40:40.966099 ignition[846]: Ignition finished successfully Sep 4 17:40:40.986016 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:40:40.995171 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 4 17:40:41.008573 ignition[876]: Ignition 2.19.0 Sep 4 17:40:41.008584 ignition[876]: Stage: fetch Sep 4 17:40:41.008798 ignition[876]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.008810 ignition[876]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.008943 ignition[876]: parsed url from cmdline: "" Sep 4 17:40:41.008947 ignition[876]: no config URL provided Sep 4 17:40:41.008953 ignition[876]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:40:41.008961 ignition[876]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:40:41.008984 ignition[876]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Sep 4 17:40:41.085440 ignition[876]: GET result: OK Sep 4 17:40:41.085552 ignition[876]: config has been read from IMDS userdata Sep 4 17:40:41.085588 ignition[876]: parsing config with SHA512: 175ceffa614c1fa967e97834cdebfcc44d8ec74e75dfc6420ce0d5edd9878075e6a8b9b1aedebe44490fa8e056f826cf36aaa8fda25752df9634e873ab0b0661 Sep 4 17:40:41.091847 unknown[876]: fetched base config from "system" Sep 4 17:40:41.091863 unknown[876]: fetched base config from "system" Sep 4 17:40:41.092249 ignition[876]: fetch: fetch complete Sep 4 17:40:41.091870 unknown[876]: fetched user config from "azure" Sep 4 17:40:41.092254 ignition[876]: fetch: fetch passed Sep 4 17:40:41.094131 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 4 17:40:41.092292 ignition[876]: Ignition finished successfully Sep 4 17:40:41.105128 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 4 17:40:41.119314 ignition[882]: Ignition 2.19.0 Sep 4 17:40:41.119325 ignition[882]: Stage: kargs Sep 4 17:40:41.119553 ignition[882]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.119565 ignition[882]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.120385 ignition[882]: kargs: kargs passed Sep 4 17:40:41.120424 ignition[882]: Ignition finished successfully Sep 4 17:40:41.128320 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 4 17:40:41.144066 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 4 17:40:41.158660 ignition[888]: Ignition 2.19.0 Sep 4 17:40:41.158674 ignition[888]: Stage: disks Sep 4 17:40:41.160775 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 4 17:40:41.158929 ignition[888]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.164173 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 4 17:40:41.158940 ignition[888]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.168267 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 17:40:41.159911 ignition[888]: disks: disks passed Sep 4 17:40:41.171734 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:40:41.159954 ignition[888]: Ignition finished successfully Sep 4 17:40:41.174517 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:40:41.177337 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:40:41.192107 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 4 17:40:41.268724 systemd-fsck[896]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Sep 4 17:40:41.273336 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 4 17:40:41.286449 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 4 17:40:41.330148 systemd-networkd[867]: enP51082s1: Gained IPv6LL Sep 4 17:40:41.384908 kernel: EXT4-fs (sda9): mounted filesystem bdbe0f61-2675-40b7-b9ae-5653402e9b23 r/w with ordered data mode. Quota mode: none. Sep 4 17:40:41.385696 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 4 17:40:41.390982 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 4 17:40:41.439116 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:40:41.446263 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 4 17:40:41.460719 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (908) Sep 4 17:40:41.460760 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:41.460782 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:41.463717 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:41.469908 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:41.471070 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 4 17:40:41.478036 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 4 17:40:41.478080 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:40:41.482706 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:40:41.489360 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 4 17:40:41.504048 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 4 17:40:41.523140 systemd-networkd[867]: eth0: Gained IPv6LL Sep 4 17:40:42.090682 coreos-metadata[910]: Sep 04 17:40:42.090 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Sep 4 17:40:42.097443 coreos-metadata[910]: Sep 04 17:40:42.093 INFO Fetch successful Sep 4 17:40:42.097443 coreos-metadata[910]: Sep 04 17:40:42.093 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Sep 4 17:40:42.105847 coreos-metadata[910]: Sep 04 17:40:42.105 INFO Fetch successful Sep 4 17:40:42.120593 coreos-metadata[910]: Sep 04 17:40:42.120 INFO wrote hostname ci-4054.1.0-a-b8703b3246 to /sysroot/etc/hostname Sep 4 17:40:42.122547 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 4 17:40:42.155377 initrd-setup-root[937]: cut: /sysroot/etc/passwd: No such file or directory Sep 4 17:40:42.207928 initrd-setup-root[944]: cut: /sysroot/etc/group: No such file or directory Sep 4 17:40:42.215051 initrd-setup-root[951]: cut: /sysroot/etc/shadow: No such file or directory Sep 4 17:40:42.221498 initrd-setup-root[958]: cut: /sysroot/etc/gshadow: No such file or directory Sep 4 17:40:43.214080 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 4 17:40:43.224970 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 4 17:40:43.232056 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 4 17:40:43.239042 kernel: BTRFS info (device sda6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:43.242364 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 4 17:40:43.264632 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 4 17:40:43.271835 ignition[1028]: INFO : Ignition 2.19.0 Sep 4 17:40:43.271835 ignition[1028]: INFO : Stage: mount Sep 4 17:40:43.279132 ignition[1028]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:43.279132 ignition[1028]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:43.279132 ignition[1028]: INFO : mount: mount passed Sep 4 17:40:43.279132 ignition[1028]: INFO : Ignition finished successfully Sep 4 17:40:43.274049 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 4 17:40:43.282992 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 4 17:40:43.300818 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:40:43.322903 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (1039) Sep 4 17:40:43.327900 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:43.327936 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:43.332509 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:43.343945 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:43.345489 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:40:43.368594 ignition[1056]: INFO : Ignition 2.19.0 Sep 4 17:40:43.368594 ignition[1056]: INFO : Stage: files Sep 4 17:40:43.373007 ignition[1056]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:43.373007 ignition[1056]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:43.373007 ignition[1056]: DEBUG : files: compiled without relabeling support, skipping Sep 4 17:40:43.373007 ignition[1056]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 4 17:40:43.373007 ignition[1056]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 4 17:40:43.437333 ignition[1056]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 4 17:40:43.442208 ignition[1056]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 4 17:40:43.442208 ignition[1056]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 4 17:40:43.437863 unknown[1056]: wrote ssh authorized keys file for user: core Sep 4 17:40:43.512492 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:40:43.519588 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 4 17:40:43.557630 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 4 17:40:43.608016 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:40:43.614257 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 4 17:40:43.619325 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 4 17:40:43.619325 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:40:43.630940 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:40:43.635508 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:40:43.640237 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Sep 4 17:40:44.174387 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 4 17:40:44.466827 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:44.466827 ignition[1056]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 4 17:40:44.484015 ignition[1056]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: files passed Sep 4 17:40:44.489202 ignition[1056]: INFO : Ignition finished successfully Sep 4 17:40:44.485823 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 4 17:40:44.531351 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 4 17:40:44.538278 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 4 17:40:44.541713 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 4 17:40:44.541796 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 4 17:40:44.573286 initrd-setup-root-after-ignition[1084]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.573286 initrd-setup-root-after-ignition[1084]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.581630 initrd-setup-root-after-ignition[1088]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.587341 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:40:44.587915 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 4 17:40:44.600125 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 4 17:40:44.625261 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 4 17:40:44.625372 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 4 17:40:44.631430 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 4 17:40:44.637405 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 4 17:40:44.640252 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 4 17:40:44.641058 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 4 17:40:44.661236 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:40:44.670329 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 4 17:40:44.682462 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:40:44.686302 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:40:44.695388 systemd[1]: Stopped target timers.target - Timer Units. Sep 4 17:40:44.700306 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 4 17:40:44.700478 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:40:44.707011 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 4 17:40:44.712263 systemd[1]: Stopped target basic.target - Basic System. Sep 4 17:40:44.719872 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 4 17:40:44.722964 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:40:44.731877 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 4 17:40:44.732100 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 4 17:40:44.732526 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:40:44.732985 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 4 17:40:44.733411 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 4 17:40:44.733846 systemd[1]: Stopped target swap.target - Swaps. Sep 4 17:40:44.734250 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 4 17:40:44.734383 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:40:44.735158 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:40:44.735616 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:40:44.736443 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 4 17:40:44.761581 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:40:44.764950 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 4 17:40:44.765088 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 4 17:40:44.771129 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 4 17:40:44.771275 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:40:44.777169 systemd[1]: ignition-files.service: Deactivated successfully. Sep 4 17:40:44.777313 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 4 17:40:44.782266 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 4 17:40:44.794340 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 4 17:40:44.835421 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 4 17:40:44.840065 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 4 17:40:44.845693 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 4 17:40:44.846071 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:40:44.858042 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 4 17:40:44.858192 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:40:44.866740 ignition[1108]: INFO : Ignition 2.19.0 Sep 4 17:40:44.866740 ignition[1108]: INFO : Stage: umount Sep 4 17:40:44.872775 ignition[1108]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:44.872775 ignition[1108]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:44.872775 ignition[1108]: INFO : umount: umount passed Sep 4 17:40:44.872775 ignition[1108]: INFO : Ignition finished successfully Sep 4 17:40:44.869815 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 4 17:40:44.869915 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 4 17:40:44.876388 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 4 17:40:44.876495 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 4 17:40:44.881560 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 4 17:40:44.881609 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 4 17:40:44.892388 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 4 17:40:44.892439 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 4 17:40:44.901148 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 4 17:40:44.901200 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 4 17:40:44.906540 systemd[1]: Stopped target network.target - Network. Sep 4 17:40:44.911172 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 4 17:40:44.911229 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:40:44.914343 systemd[1]: Stopped target paths.target - Path Units. Sep 4 17:40:44.914437 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 4 17:40:44.919284 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:40:44.919370 systemd[1]: Stopped target slices.target - Slice Units. Sep 4 17:40:44.919809 systemd[1]: Stopped target sockets.target - Socket Units. Sep 4 17:40:44.920283 systemd[1]: iscsid.socket: Deactivated successfully. Sep 4 17:40:44.920325 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:40:44.920707 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 4 17:40:44.920739 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:40:44.921567 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 4 17:40:44.921607 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 4 17:40:44.922025 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 4 17:40:44.922059 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 4 17:40:44.922588 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 4 17:40:44.923065 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 4 17:40:44.950950 systemd-networkd[867]: eth0: DHCPv6 lease lost Sep 4 17:40:44.952665 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 4 17:40:44.952770 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 4 17:40:44.968548 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 4 17:40:44.969146 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 4 17:40:44.969270 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 4 17:40:44.974848 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 4 17:40:44.974943 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:40:44.993041 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 4 17:40:44.999191 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 4 17:40:44.999250 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:40:45.005133 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 17:40:45.005180 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:40:45.005275 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 4 17:40:45.005312 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 4 17:40:45.064599 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 4 17:40:45.064671 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:40:45.074053 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:40:45.091470 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 4 17:40:45.091628 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:40:45.098730 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 4 17:40:45.098769 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 4 17:40:45.104867 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 4 17:40:45.104927 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:40:45.121303 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 4 17:40:45.121368 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:40:45.133524 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 4 17:40:45.136045 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: Data path switched from VF: enP51082s1 Sep 4 17:40:45.133592 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 4 17:40:45.139690 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:40:45.139734 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:45.154094 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 4 17:40:45.160014 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 4 17:40:45.160074 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:40:45.163410 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 4 17:40:45.163456 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:40:45.179125 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 4 17:40:45.179178 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:40:45.187857 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:45.187928 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:45.196454 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 4 17:40:45.196558 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 4 17:40:45.204566 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 4 17:40:45.204677 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 4 17:40:45.444074 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 4 17:40:45.444204 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 4 17:40:45.452473 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 4 17:40:45.456947 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 4 17:40:45.457014 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 4 17:40:45.468100 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 4 17:40:45.475119 systemd[1]: Switching root. Sep 4 17:40:45.567469 systemd-journald[176]: Journal stopped Sep 4 17:40:35.094196 kernel: Linux version 6.6.48-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Wed Sep 4 15:54:07 -00 2024 Sep 4 17:40:35.094225 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.094236 kernel: BIOS-provided physical RAM map: Sep 4 17:40:35.094242 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 4 17:40:35.094249 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Sep 4 17:40:35.094257 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ff40fff] usable Sep 4 17:40:35.094264 kernel: BIOS-e820: [mem 0x000000003ff41000-0x000000003ff70fff] type 20 Sep 4 17:40:35.094276 kernel: BIOS-e820: [mem 0x000000003ff71000-0x000000003ffc8fff] reserved Sep 4 17:40:35.094282 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Sep 4 17:40:35.094289 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Sep 4 17:40:35.094298 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Sep 4 17:40:35.094304 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Sep 4 17:40:35.094311 kernel: printk: bootconsole [earlyser0] enabled Sep 4 17:40:35.094320 kernel: NX (Execute Disable) protection: active Sep 4 17:40:35.094330 kernel: APIC: Static calls initialized Sep 4 17:40:35.094340 kernel: efi: EFI v2.7 by Microsoft Sep 4 17:40:35.094348 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff85000 SMBIOS 3.0=0x3ff83000 MEMATTR=0x3f5c1a98 Sep 4 17:40:35.094355 kernel: SMBIOS 3.1.0 present. Sep 4 17:40:35.094365 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 03/08/2024 Sep 4 17:40:35.094372 kernel: Hypervisor detected: Microsoft Hyper-V Sep 4 17:40:35.094380 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0x64e24, misc 0xbed7b2 Sep 4 17:40:35.094389 kernel: Hyper-V: Host Build 10.0.20348.1633-1-0 Sep 4 17:40:35.094396 kernel: Hyper-V: Nested features: 0x1e0101 Sep 4 17:40:35.094406 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Sep 4 17:40:35.094416 kernel: Hyper-V: Using hypercall for remote TLB flush Sep 4 17:40:35.094426 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Sep 4 17:40:35.094434 kernel: clocksource: hyperv_clocksource_msr: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Sep 4 17:40:35.094442 kernel: tsc: Marking TSC unstable due to running on Hyper-V Sep 4 17:40:35.094452 kernel: tsc: Detected 2593.907 MHz processor Sep 4 17:40:35.094459 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 4 17:40:35.094469 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 4 17:40:35.094478 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x400000000 Sep 4 17:40:35.094486 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 4 17:40:35.094497 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 4 17:40:35.094504 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved Sep 4 17:40:35.094513 kernel: last_pfn = 0x40000 max_arch_pfn = 0x400000000 Sep 4 17:40:35.094521 kernel: Using GB pages for direct mapping Sep 4 17:40:35.094528 kernel: Secure boot disabled Sep 4 17:40:35.094544 kernel: ACPI: Early table checksum verification disabled Sep 4 17:40:35.094554 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Sep 4 17:40:35.094569 kernel: ACPI: XSDT 0x000000003FFF90E8 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094581 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094590 kernel: ACPI: DSDT 0x000000003FFD6000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Sep 4 17:40:35.094599 kernel: ACPI: FACS 0x000000003FFFE000 000040 Sep 4 17:40:35.094610 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094618 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094629 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094639 kernel: ACPI: APIC 0x000000003FFD5000 000058 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094649 kernel: ACPI: SRAT 0x000000003FFD4000 0002D0 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094657 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094665 kernel: ACPI: FPDT 0x000000003FFD2000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Sep 4 17:40:35.094676 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Sep 4 17:40:35.094683 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4183] Sep 4 17:40:35.094694 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Sep 4 17:40:35.094702 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Sep 4 17:40:35.094714 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Sep 4 17:40:35.094722 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Sep 4 17:40:35.094730 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5057] Sep 4 17:40:35.094741 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd42cf] Sep 4 17:40:35.094748 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Sep 4 17:40:35.094759 kernel: ACPI: Reserving FPDT table memory at [mem 0x3ffd2000-0x3ffd2033] Sep 4 17:40:35.094767 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 4 17:40:35.094775 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 4 17:40:35.094784 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Sep 4 17:40:35.094795 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] hotplug Sep 4 17:40:35.094805 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2c0000000-0xfdfffffff] hotplug Sep 4 17:40:35.094812 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Sep 4 17:40:35.094823 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Sep 4 17:40:35.094831 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Sep 4 17:40:35.094840 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Sep 4 17:40:35.094849 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Sep 4 17:40:35.094856 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Sep 4 17:40:35.094867 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Sep 4 17:40:35.094877 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Sep 4 17:40:35.094888 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Sep 4 17:40:35.094896 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000000-0x1ffffffffffff] hotplug Sep 4 17:40:35.094904 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2000000000000-0x3ffffffffffff] hotplug Sep 4 17:40:35.094914 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x4000000000000-0x7ffffffffffff] hotplug Sep 4 17:40:35.094921 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x8000000000000-0xfffffffffffff] hotplug Sep 4 17:40:35.094937 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00000000-0x2bfffffff] Sep 4 17:40:35.094948 kernel: NODE_DATA(0) allocated [mem 0x2bfffa000-0x2bfffffff] Sep 4 17:40:35.094956 kernel: Zone ranges: Sep 4 17:40:35.094969 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 4 17:40:35.094979 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Sep 4 17:40:35.094989 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Sep 4 17:40:35.094999 kernel: Movable zone start for each node Sep 4 17:40:35.095007 kernel: Early memory node ranges Sep 4 17:40:35.095015 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 4 17:40:35.095025 kernel: node 0: [mem 0x0000000000100000-0x000000003ff40fff] Sep 4 17:40:35.095033 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Sep 4 17:40:35.095043 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Sep 4 17:40:35.095053 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Sep 4 17:40:35.095064 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 4 17:40:35.095072 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 4 17:40:35.095089 kernel: On node 0, zone DMA32: 190 pages in unavailable ranges Sep 4 17:40:35.095097 kernel: ACPI: PM-Timer IO Port: 0x408 Sep 4 17:40:35.095108 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) Sep 4 17:40:35.095115 kernel: IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 Sep 4 17:40:35.095124 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 4 17:40:35.095134 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 4 17:40:35.095145 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Sep 4 17:40:35.095154 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 4 17:40:35.095162 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Sep 4 17:40:35.095172 kernel: Booting paravirtualized kernel on Hyper-V Sep 4 17:40:35.095180 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 4 17:40:35.095190 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 4 17:40:35.095199 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u1048576 Sep 4 17:40:35.095206 kernel: pcpu-alloc: s196904 r8192 d32472 u1048576 alloc=1*2097152 Sep 4 17:40:35.095217 kernel: pcpu-alloc: [0] 0 1 Sep 4 17:40:35.095226 kernel: Hyper-V: PV spinlocks enabled Sep 4 17:40:35.095237 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 4 17:40:35.095246 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.095257 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 4 17:40:35.095265 kernel: random: crng init done Sep 4 17:40:35.095273 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Sep 4 17:40:35.095283 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 4 17:40:35.095290 kernel: Fallback order for Node 0: 0 Sep 4 17:40:35.095303 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2062618 Sep 4 17:40:35.095326 kernel: Policy zone: Normal Sep 4 17:40:35.095340 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 4 17:40:35.095350 kernel: software IO TLB: area num 2. Sep 4 17:40:35.095360 kernel: Memory: 8077076K/8387460K available (12288K kernel code, 2304K rwdata, 22708K rodata, 42704K init, 2488K bss, 310124K reserved, 0K cma-reserved) Sep 4 17:40:35.095371 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 4 17:40:35.095381 kernel: ftrace: allocating 37748 entries in 148 pages Sep 4 17:40:35.095392 kernel: ftrace: allocated 148 pages with 3 groups Sep 4 17:40:35.095400 kernel: Dynamic Preempt: voluntary Sep 4 17:40:35.095410 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 4 17:40:35.095420 kernel: rcu: RCU event tracing is enabled. Sep 4 17:40:35.095433 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 4 17:40:35.095443 kernel: Trampoline variant of Tasks RCU enabled. Sep 4 17:40:35.095452 kernel: Rude variant of Tasks RCU enabled. Sep 4 17:40:35.095462 kernel: Tracing variant of Tasks RCU enabled. Sep 4 17:40:35.095470 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 4 17:40:35.095483 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 4 17:40:35.095492 kernel: Using NULL legacy PIC Sep 4 17:40:35.095503 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Sep 4 17:40:35.095511 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 4 17:40:35.095522 kernel: Console: colour dummy device 80x25 Sep 4 17:40:35.095530 kernel: printk: console [tty1] enabled Sep 4 17:40:35.095540 kernel: printk: console [ttyS0] enabled Sep 4 17:40:35.095549 kernel: printk: bootconsole [earlyser0] disabled Sep 4 17:40:35.095558 kernel: ACPI: Core revision 20230628 Sep 4 17:40:35.095569 kernel: Failed to register legacy timer interrupt Sep 4 17:40:35.095579 kernel: APIC: Switch to symmetric I/O mode setup Sep 4 17:40:35.095590 kernel: Hyper-V: enabling crash_kexec_post_notifiers Sep 4 17:40:35.095598 kernel: Hyper-V: Using IPI hypercalls Sep 4 17:40:35.099814 kernel: APIC: send_IPI() replaced with hv_send_ipi() Sep 4 17:40:35.099840 kernel: APIC: send_IPI_mask() replaced with hv_send_ipi_mask() Sep 4 17:40:35.099854 kernel: APIC: send_IPI_mask_allbutself() replaced with hv_send_ipi_mask_allbutself() Sep 4 17:40:35.099869 kernel: APIC: send_IPI_allbutself() replaced with hv_send_ipi_allbutself() Sep 4 17:40:35.099884 kernel: APIC: send_IPI_all() replaced with hv_send_ipi_all() Sep 4 17:40:35.099898 kernel: APIC: send_IPI_self() replaced with hv_send_ipi_self() Sep 4 17:40:35.099917 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 5187.81 BogoMIPS (lpj=2593907) Sep 4 17:40:35.099932 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 4 17:40:35.099946 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Sep 4 17:40:35.099960 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 4 17:40:35.099973 kernel: Spectre V2 : Mitigation: Retpolines Sep 4 17:40:35.099986 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Sep 4 17:40:35.100000 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Sep 4 17:40:35.100012 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Sep 4 17:40:35.100029 kernel: RETBleed: Vulnerable Sep 4 17:40:35.100053 kernel: Speculative Store Bypass: Vulnerable Sep 4 17:40:35.100066 kernel: TAA: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:40:35.100087 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:40:35.100101 kernel: GDS: Unknown: Dependent on hypervisor status Sep 4 17:40:35.100115 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 4 17:40:35.100127 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 4 17:40:35.100140 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 4 17:40:35.100154 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Sep 4 17:40:35.100168 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Sep 4 17:40:35.100182 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Sep 4 17:40:35.100196 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 4 17:40:35.100214 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Sep 4 17:40:35.100229 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Sep 4 17:40:35.100244 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Sep 4 17:40:35.100260 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Sep 4 17:40:35.100275 kernel: Freeing SMP alternatives memory: 32K Sep 4 17:40:35.100290 kernel: pid_max: default: 32768 minimum: 301 Sep 4 17:40:35.100304 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 4 17:40:35.100319 kernel: landlock: Up and running. Sep 4 17:40:35.100334 kernel: SELinux: Initializing. Sep 4 17:40:35.100349 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.100363 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.100378 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x7) Sep 4 17:40:35.100396 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100411 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100427 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:40:35.100442 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Sep 4 17:40:35.100457 kernel: signal: max sigframe size: 3632 Sep 4 17:40:35.100472 kernel: rcu: Hierarchical SRCU implementation. Sep 4 17:40:35.100487 kernel: rcu: Max phase no-delay instances is 400. Sep 4 17:40:35.100503 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 4 17:40:35.100518 kernel: smp: Bringing up secondary CPUs ... Sep 4 17:40:35.100536 kernel: smpboot: x86: Booting SMP configuration: Sep 4 17:40:35.100551 kernel: .... node #0, CPUs: #1 Sep 4 17:40:35.100564 kernel: TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. Sep 4 17:40:35.100578 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 4 17:40:35.100593 kernel: smp: Brought up 1 node, 2 CPUs Sep 4 17:40:35.100607 kernel: smpboot: Max logical packages: 1 Sep 4 17:40:35.100622 kernel: smpboot: Total of 2 processors activated (10375.62 BogoMIPS) Sep 4 17:40:35.100637 kernel: devtmpfs: initialized Sep 4 17:40:35.100654 kernel: x86/mm: Memory block size: 128MB Sep 4 17:40:35.100669 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Sep 4 17:40:35.100684 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 4 17:40:35.100699 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 4 17:40:35.100713 kernel: pinctrl core: initialized pinctrl subsystem Sep 4 17:40:35.100727 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 4 17:40:35.100741 kernel: audit: initializing netlink subsys (disabled) Sep 4 17:40:35.100755 kernel: audit: type=2000 audit(1725471634.028:1): state=initialized audit_enabled=0 res=1 Sep 4 17:40:35.100769 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 4 17:40:35.100786 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 4 17:40:35.100800 kernel: cpuidle: using governor menu Sep 4 17:40:35.100814 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 4 17:40:35.100828 kernel: dca service started, version 1.12.1 Sep 4 17:40:35.100842 kernel: e820: reserve RAM buffer [mem 0x3ff41000-0x3fffffff] Sep 4 17:40:35.100857 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 4 17:40:35.100871 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 4 17:40:35.100885 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 4 17:40:35.100899 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 4 17:40:35.100915 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 4 17:40:35.100930 kernel: ACPI: Added _OSI(Module Device) Sep 4 17:40:35.100944 kernel: ACPI: Added _OSI(Processor Device) Sep 4 17:40:35.100958 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Sep 4 17:40:35.100972 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 4 17:40:35.100986 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 4 17:40:35.101000 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 4 17:40:35.101014 kernel: ACPI: Interpreter enabled Sep 4 17:40:35.101028 kernel: ACPI: PM: (supports S0 S5) Sep 4 17:40:35.101045 kernel: ACPI: Using IOAPIC for interrupt routing Sep 4 17:40:35.101059 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 4 17:40:35.101073 kernel: PCI: Ignoring E820 reservations for host bridge windows Sep 4 17:40:35.101104 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Sep 4 17:40:35.101118 kernel: iommu: Default domain type: Translated Sep 4 17:40:35.101133 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 4 17:40:35.101147 kernel: efivars: Registered efivars operations Sep 4 17:40:35.101160 kernel: PCI: Using ACPI for IRQ routing Sep 4 17:40:35.101175 kernel: PCI: System does not support PCI Sep 4 17:40:35.101191 kernel: vgaarb: loaded Sep 4 17:40:35.101205 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page Sep 4 17:40:35.101219 kernel: VFS: Disk quotas dquot_6.6.0 Sep 4 17:40:35.101233 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 4 17:40:35.101248 kernel: pnp: PnP ACPI init Sep 4 17:40:35.101261 kernel: pnp: PnP ACPI: found 3 devices Sep 4 17:40:35.101275 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 4 17:40:35.101290 kernel: NET: Registered PF_INET protocol family Sep 4 17:40:35.101304 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 4 17:40:35.101320 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Sep 4 17:40:35.101335 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 4 17:40:35.101348 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 4 17:40:35.101363 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Sep 4 17:40:35.101377 kernel: TCP: Hash tables configured (established 65536 bind 65536) Sep 4 17:40:35.101391 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.101405 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 4 17:40:35.101419 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 4 17:40:35.101433 kernel: NET: Registered PF_XDP protocol family Sep 4 17:40:35.101449 kernel: PCI: CLS 0 bytes, default 64 Sep 4 17:40:35.101463 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 4 17:40:35.101476 kernel: software IO TLB: mapped [mem 0x000000003b5c1000-0x000000003f5c1000] (64MB) Sep 4 17:40:35.101488 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 4 17:40:35.101503 kernel: Initialise system trusted keyrings Sep 4 17:40:35.101516 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Sep 4 17:40:35.101530 kernel: Key type asymmetric registered Sep 4 17:40:35.101543 kernel: Asymmetric key parser 'x509' registered Sep 4 17:40:35.101572 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 4 17:40:35.101589 kernel: io scheduler mq-deadline registered Sep 4 17:40:35.101601 kernel: io scheduler kyber registered Sep 4 17:40:35.101615 kernel: io scheduler bfq registered Sep 4 17:40:35.101630 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 4 17:40:35.101647 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 4 17:40:35.101661 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 4 17:40:35.101675 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Sep 4 17:40:35.101689 kernel: i8042: PNP: No PS/2 controller found. Sep 4 17:40:35.101865 kernel: rtc_cmos 00:02: registered as rtc0 Sep 4 17:40:35.101994 kernel: rtc_cmos 00:02: setting system clock to 2024-09-04T17:40:34 UTC (1725471634) Sep 4 17:40:35.102136 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Sep 4 17:40:35.102153 kernel: intel_pstate: CPU model not supported Sep 4 17:40:35.102167 kernel: efifb: probing for efifb Sep 4 17:40:35.102181 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Sep 4 17:40:35.102195 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Sep 4 17:40:35.102209 kernel: efifb: scrolling: redraw Sep 4 17:40:35.102223 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 4 17:40:35.102240 kernel: Console: switching to colour frame buffer device 128x48 Sep 4 17:40:35.102254 kernel: fb0: EFI VGA frame buffer device Sep 4 17:40:35.102268 kernel: pstore: Using crash dump compression: deflate Sep 4 17:40:35.102282 kernel: pstore: Registered efi_pstore as persistent store backend Sep 4 17:40:35.102296 kernel: NET: Registered PF_INET6 protocol family Sep 4 17:40:35.102309 kernel: Segment Routing with IPv6 Sep 4 17:40:35.102323 kernel: In-situ OAM (IOAM) with IPv6 Sep 4 17:40:35.102336 kernel: NET: Registered PF_PACKET protocol family Sep 4 17:40:35.102350 kernel: Key type dns_resolver registered Sep 4 17:40:35.102366 kernel: IPI shorthand broadcast: enabled Sep 4 17:40:35.102390 kernel: sched_clock: Marking stable (851002800, 50107800)->(1128054500, -226943900) Sep 4 17:40:35.102403 kernel: registered taskstats version 1 Sep 4 17:40:35.102417 kernel: Loading compiled-in X.509 certificates Sep 4 17:40:35.102429 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.48-flatcar: 8669771ab5e11f458b79e6634fe685dacc266b18' Sep 4 17:40:35.102442 kernel: Key type .fscrypt registered Sep 4 17:40:35.102455 kernel: Key type fscrypt-provisioning registered Sep 4 17:40:35.102467 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 4 17:40:35.102483 kernel: ima: Allocated hash algorithm: sha1 Sep 4 17:40:35.102496 kernel: ima: No architecture policies found Sep 4 17:40:35.102509 kernel: clk: Disabling unused clocks Sep 4 17:40:35.102522 kernel: Freeing unused kernel image (initmem) memory: 42704K Sep 4 17:40:35.102535 kernel: Write protecting the kernel read-only data: 36864k Sep 4 17:40:35.102548 kernel: Freeing unused kernel image (rodata/data gap) memory: 1868K Sep 4 17:40:35.102561 kernel: Run /init as init process Sep 4 17:40:35.102574 kernel: with arguments: Sep 4 17:40:35.102587 kernel: /init Sep 4 17:40:35.102599 kernel: with environment: Sep 4 17:40:35.102614 kernel: HOME=/ Sep 4 17:40:35.102630 kernel: TERM=linux Sep 4 17:40:35.102643 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 4 17:40:35.102657 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:40:35.102673 systemd[1]: Detected virtualization microsoft. Sep 4 17:40:35.102686 systemd[1]: Detected architecture x86-64. Sep 4 17:40:35.102699 systemd[1]: Running in initrd. Sep 4 17:40:35.102722 systemd[1]: No hostname configured, using default hostname. Sep 4 17:40:35.102735 systemd[1]: Hostname set to . Sep 4 17:40:35.102749 systemd[1]: Initializing machine ID from random generator. Sep 4 17:40:35.102763 systemd[1]: Queued start job for default target initrd.target. Sep 4 17:40:35.102780 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:40:35.102794 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:40:35.102808 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 4 17:40:35.102824 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:40:35.102842 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 4 17:40:35.102857 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 4 17:40:35.102874 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 4 17:40:35.102889 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 4 17:40:35.102904 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:40:35.102918 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:40:35.102933 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:40:35.102950 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:40:35.102965 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:40:35.102979 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:40:35.102994 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:40:35.103008 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:40:35.103022 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 17:40:35.103037 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 4 17:40:35.103052 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:40:35.103066 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:40:35.103699 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:40:35.103718 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:40:35.103734 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 4 17:40:35.103751 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:40:35.103770 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 4 17:40:35.103785 systemd[1]: Starting systemd-fsck-usr.service... Sep 4 17:40:35.103799 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:40:35.103814 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:40:35.103865 systemd-journald[176]: Collecting audit messages is disabled. Sep 4 17:40:35.103898 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:35.103913 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 4 17:40:35.103945 systemd-journald[176]: Journal started Sep 4 17:40:35.103985 systemd-journald[176]: Runtime Journal (/run/log/journal/ee2e414552134498a0bdd253f0c95dcc) is 8.0M, max 158.8M, 150.8M free. Sep 4 17:40:35.100067 systemd-modules-load[177]: Inserted module 'overlay' Sep 4 17:40:35.115934 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:40:35.119340 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:40:35.122786 systemd[1]: Finished systemd-fsck-usr.service. Sep 4 17:40:35.131778 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:35.144106 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 4 17:40:35.146260 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:35.152068 kernel: Bridge firewalling registered Sep 4 17:40:35.148597 systemd-modules-load[177]: Inserted module 'br_netfilter' Sep 4 17:40:35.162246 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:40:35.171822 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 17:40:35.175190 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:40:35.184973 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:35.188855 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:40:35.192516 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:40:35.212228 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 4 17:40:35.222269 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:40:35.228068 dracut-cmdline[204]: dracut-dracut-053 Sep 4 17:40:35.229951 dracut-cmdline[204]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:40:35.249374 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:40:35.262691 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:40:35.277521 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:40:35.283639 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:40:35.330097 kernel: SCSI subsystem initialized Sep 4 17:40:35.330619 systemd-resolved[260]: Positive Trust Anchors: Sep 4 17:40:35.330635 systemd-resolved[260]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:40:35.330683 systemd-resolved[260]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 17:40:35.357050 systemd-resolved[260]: Defaulting to hostname 'linux'. Sep 4 17:40:35.365241 kernel: Loading iSCSI transport class v2.0-870. Sep 4 17:40:35.358253 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:40:35.368057 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:40:35.387101 kernel: iscsi: registered transport (tcp) Sep 4 17:40:35.408086 kernel: iscsi: registered transport (qla4xxx) Sep 4 17:40:35.408154 kernel: QLogic iSCSI HBA Driver Sep 4 17:40:35.444621 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 4 17:40:35.456333 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 4 17:40:35.486719 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 4 17:40:35.486790 kernel: device-mapper: uevent: version 1.0.3 Sep 4 17:40:35.490054 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 4 17:40:35.530101 kernel: raid6: avx512x4 gen() 18093 MB/s Sep 4 17:40:35.549090 kernel: raid6: avx512x2 gen() 18049 MB/s Sep 4 17:40:35.568087 kernel: raid6: avx512x1 gen() 17920 MB/s Sep 4 17:40:35.587092 kernel: raid6: avx2x4 gen() 18120 MB/s Sep 4 17:40:35.605088 kernel: raid6: avx2x2 gen() 17980 MB/s Sep 4 17:40:35.625548 kernel: raid6: avx2x1 gen() 13880 MB/s Sep 4 17:40:35.625607 kernel: raid6: using algorithm avx2x4 gen() 18120 MB/s Sep 4 17:40:35.648197 kernel: raid6: .... xor() 6986 MB/s, rmw enabled Sep 4 17:40:35.648234 kernel: raid6: using avx512x2 recovery algorithm Sep 4 17:40:35.670102 kernel: xor: automatically using best checksumming function avx Sep 4 17:40:35.822104 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 4 17:40:35.831946 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:40:35.841233 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:40:35.853435 systemd-udevd[396]: Using default interface naming scheme 'v255'. Sep 4 17:40:35.857899 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:40:35.873206 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 4 17:40:35.884630 dracut-pre-trigger[406]: rd.md=0: removing MD RAID activation Sep 4 17:40:35.909714 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:40:35.920197 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:40:35.960636 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:40:35.974299 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 4 17:40:36.000406 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 4 17:40:36.010595 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:40:36.010680 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:40:36.011591 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:40:36.034471 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 4 17:40:36.058718 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:40:36.071519 kernel: cryptd: max_cpu_qlen set to 1000 Sep 4 17:40:36.082851 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:40:36.086201 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:36.096120 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:36.108370 kernel: AVX2 version of gcm_enc/dec engaged. Sep 4 17:40:36.108423 kernel: AES CTR mode by8 optimization enabled Sep 4 17:40:36.105116 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:36.118219 kernel: hv_vmbus: Vmbus version:5.2 Sep 4 17:40:36.105377 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.108842 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.121558 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.129055 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:36.129155 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.150168 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 4 17:40:36.150213 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 4 17:40:36.154119 kernel: hv_vmbus: registering driver hyperv_keyboard Sep 4 17:40:36.155428 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:36.177264 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:36.185259 kernel: PTP clock support registered Sep 4 17:40:36.189264 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:40:37.418381 kernel: hv_utils: Registering HyperV Utility Driver Sep 4 17:40:37.418400 kernel: hv_vmbus: registering driver hv_utils Sep 4 17:40:37.418414 kernel: hv_utils: Heartbeat IC version 3.0 Sep 4 17:40:37.418425 kernel: hv_utils: Shutdown IC version 3.2 Sep 4 17:40:37.418438 kernel: hv_utils: TimeSync IC version 4.0 Sep 4 17:40:37.418449 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Sep 4 17:40:37.418460 kernel: hv_vmbus: registering driver hv_storvsc Sep 4 17:40:37.418479 kernel: scsi host1: storvsc_host_t Sep 4 17:40:37.418630 kernel: scsi host0: storvsc_host_t Sep 4 17:40:37.412166 systemd-resolved[260]: Clock change detected. Flushing caches. Sep 4 17:40:37.438117 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Sep 4 17:40:37.445455 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Sep 4 17:40:37.445498 kernel: hv_vmbus: registering driver hv_netvsc Sep 4 17:40:37.449099 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 4 17:40:37.461934 kernel: hv_vmbus: registering driver hid_hyperv Sep 4 17:40:37.469391 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Sep 4 17:40:37.469429 kernel: hid 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Sep 4 17:40:37.476703 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:37.497092 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Sep 4 17:40:37.497317 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Sep 4 17:40:37.503903 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Sep 4 17:40:37.517101 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Sep 4 17:40:37.517393 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Sep 4 17:40:37.517909 kernel: sd 0:0:0:0: [sda] Write Protect is off Sep 4 17:40:37.523133 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Sep 4 17:40:37.523398 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Sep 4 17:40:37.529904 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:37.529939 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Sep 4 17:40:37.655999 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: VF slot 1 added Sep 4 17:40:37.663929 kernel: hv_vmbus: registering driver hv_pci Sep 4 17:40:37.667922 kernel: hv_pci 879eec16-c78a-4323-af28-425ed18503dc: PCI VMBus probing: Using version 0x10004 Sep 4 17:40:37.672919 kernel: hv_pci 879eec16-c78a-4323-af28-425ed18503dc: PCI host bridge to bus c78a:00 Sep 4 17:40:37.673079 kernel: pci_bus c78a:00: root bus resource [mem 0xfe0000000-0xfe00fffff window] Sep 4 17:40:37.678078 kernel: pci_bus c78a:00: No busn resource found for root bus, will use [bus 00-ff] Sep 4 17:40:37.683212 kernel: pci c78a:00:02.0: [15b3:1016] type 00 class 0x020000 Sep 4 17:40:37.686897 kernel: pci c78a:00:02.0: reg 0x10: [mem 0xfe0000000-0xfe00fffff 64bit pref] Sep 4 17:40:37.690947 kernel: pci c78a:00:02.0: enabling Extended Tags Sep 4 17:40:37.701920 kernel: pci c78a:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at c78a:00:02.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Sep 4 17:40:37.708396 kernel: pci_bus c78a:00: busn_res: [bus 00-ff] end is updated to 00 Sep 4 17:40:37.708648 kernel: pci c78a:00:02.0: BAR 0: assigned [mem 0xfe0000000-0xfe00fffff 64bit pref] Sep 4 17:40:37.883429 kernel: mlx5_core c78a:00:02.0: enabling device (0000 -> 0002) Sep 4 17:40:37.887912 kernel: mlx5_core c78a:00:02.0: firmware version: 14.30.1284 Sep 4 17:40:38.034682 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Sep 4 17:40:38.083936 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (455) Sep 4 17:40:38.100349 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Sep 4 17:40:38.121101 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: VF registering: eth1 Sep 4 17:40:38.121342 kernel: mlx5_core c78a:00:02.0 eth1: joined to eth0 Sep 4 17:40:38.125722 kernel: mlx5_core c78a:00:02.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Sep 4 17:40:38.135905 kernel: mlx5_core c78a:00:02.0 enP51082s1: renamed from eth1 Sep 4 17:40:38.147918 kernel: BTRFS: device fsid 0dc40443-7f77-4fa7-b5e4-579d4bba0772 devid 1 transid 37 /dev/sda3 scanned by (udev-worker) (450) Sep 4 17:40:38.172219 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Sep 4 17:40:38.180533 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Sep 4 17:40:38.189459 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Sep 4 17:40:38.199037 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 4 17:40:38.213903 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:38.223906 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:39.230967 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 4 17:40:39.231625 disk-uuid[597]: The operation has completed successfully. Sep 4 17:40:39.313275 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 4 17:40:39.313386 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 4 17:40:39.334063 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 4 17:40:39.340268 sh[683]: Success Sep 4 17:40:39.373155 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 4 17:40:39.595706 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 4 17:40:39.610015 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 4 17:40:39.615597 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 4 17:40:39.633902 kernel: BTRFS info (device dm-0): first mount of filesystem 0dc40443-7f77-4fa7-b5e4-579d4bba0772 Sep 4 17:40:39.633941 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:39.639803 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 4 17:40:39.642882 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 4 17:40:39.645405 kernel: BTRFS info (device dm-0): using free space tree Sep 4 17:40:39.928280 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 4 17:40:39.935986 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 4 17:40:39.945044 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 4 17:40:39.953064 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 4 17:40:39.977560 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:39.977612 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:39.977632 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:39.996906 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:40.012166 kernel: BTRFS info (device sda6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:40.011760 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 4 17:40:40.019446 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 4 17:40:40.031096 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 4 17:40:40.053374 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:40:40.066046 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:40:40.085631 systemd-networkd[867]: lo: Link UP Sep 4 17:40:40.085641 systemd-networkd[867]: lo: Gained carrier Sep 4 17:40:40.087706 systemd-networkd[867]: Enumeration completed Sep 4 17:40:40.088298 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:40:40.092252 systemd-networkd[867]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:40.092256 systemd-networkd[867]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:40:40.093017 systemd[1]: Reached target network.target - Network. Sep 4 17:40:40.159952 kernel: mlx5_core c78a:00:02.0 enP51082s1: Link up Sep 4 17:40:40.194923 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: Data path switched to VF: enP51082s1 Sep 4 17:40:40.195994 systemd-networkd[867]: enP51082s1: Link UP Sep 4 17:40:40.196160 systemd-networkd[867]: eth0: Link UP Sep 4 17:40:40.196406 systemd-networkd[867]: eth0: Gained carrier Sep 4 17:40:40.196420 systemd-networkd[867]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:40.211150 systemd-networkd[867]: enP51082s1: Gained carrier Sep 4 17:40:40.232993 systemd-networkd[867]: eth0: DHCPv4 address 10.200.4.34/24, gateway 10.200.4.1 acquired from 168.63.129.16 Sep 4 17:40:40.963967 ignition[846]: Ignition 2.19.0 Sep 4 17:40:40.963979 ignition[846]: Stage: fetch-offline Sep 4 17:40:40.964024 ignition[846]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:40.964036 ignition[846]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:40.964147 ignition[846]: parsed url from cmdline: "" Sep 4 17:40:40.964151 ignition[846]: no config URL provided Sep 4 17:40:40.964159 ignition[846]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:40:40.964170 ignition[846]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:40:40.964176 ignition[846]: failed to fetch config: resource requires networking Sep 4 17:40:40.966099 ignition[846]: Ignition finished successfully Sep 4 17:40:40.986016 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:40:40.995171 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 4 17:40:41.008573 ignition[876]: Ignition 2.19.0 Sep 4 17:40:41.008584 ignition[876]: Stage: fetch Sep 4 17:40:41.008798 ignition[876]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.008810 ignition[876]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.008943 ignition[876]: parsed url from cmdline: "" Sep 4 17:40:41.008947 ignition[876]: no config URL provided Sep 4 17:40:41.008953 ignition[876]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:40:41.008961 ignition[876]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:40:41.008984 ignition[876]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Sep 4 17:40:41.085440 ignition[876]: GET result: OK Sep 4 17:40:41.085552 ignition[876]: config has been read from IMDS userdata Sep 4 17:40:41.085588 ignition[876]: parsing config with SHA512: 175ceffa614c1fa967e97834cdebfcc44d8ec74e75dfc6420ce0d5edd9878075e6a8b9b1aedebe44490fa8e056f826cf36aaa8fda25752df9634e873ab0b0661 Sep 4 17:40:41.091847 unknown[876]: fetched base config from "system" Sep 4 17:40:41.091863 unknown[876]: fetched base config from "system" Sep 4 17:40:41.092249 ignition[876]: fetch: fetch complete Sep 4 17:40:41.091870 unknown[876]: fetched user config from "azure" Sep 4 17:40:41.092254 ignition[876]: fetch: fetch passed Sep 4 17:40:41.094131 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 4 17:40:41.092292 ignition[876]: Ignition finished successfully Sep 4 17:40:41.105128 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 4 17:40:41.119314 ignition[882]: Ignition 2.19.0 Sep 4 17:40:41.119325 ignition[882]: Stage: kargs Sep 4 17:40:41.119553 ignition[882]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.119565 ignition[882]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.120385 ignition[882]: kargs: kargs passed Sep 4 17:40:41.120424 ignition[882]: Ignition finished successfully Sep 4 17:40:41.128320 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 4 17:40:41.144066 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 4 17:40:41.158660 ignition[888]: Ignition 2.19.0 Sep 4 17:40:41.158674 ignition[888]: Stage: disks Sep 4 17:40:41.160775 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 4 17:40:41.158929 ignition[888]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:41.164173 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 4 17:40:41.158940 ignition[888]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:41.168267 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 17:40:41.159911 ignition[888]: disks: disks passed Sep 4 17:40:41.171734 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:40:41.159954 ignition[888]: Ignition finished successfully Sep 4 17:40:41.174517 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:40:41.177337 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:40:41.192107 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 4 17:40:41.268724 systemd-fsck[896]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Sep 4 17:40:41.273336 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 4 17:40:41.286449 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 4 17:40:41.330148 systemd-networkd[867]: enP51082s1: Gained IPv6LL Sep 4 17:40:41.384908 kernel: EXT4-fs (sda9): mounted filesystem bdbe0f61-2675-40b7-b9ae-5653402e9b23 r/w with ordered data mode. Quota mode: none. Sep 4 17:40:41.385696 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 4 17:40:41.390982 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 4 17:40:41.439116 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:40:41.446263 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 4 17:40:41.460719 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (908) Sep 4 17:40:41.460760 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:41.460782 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:41.463717 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:41.469908 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:41.471070 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Sep 4 17:40:41.478036 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 4 17:40:41.478080 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:40:41.482706 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:40:41.489360 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 4 17:40:41.504048 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 4 17:40:41.523140 systemd-networkd[867]: eth0: Gained IPv6LL Sep 4 17:40:42.090682 coreos-metadata[910]: Sep 04 17:40:42.090 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Sep 4 17:40:42.097443 coreos-metadata[910]: Sep 04 17:40:42.093 INFO Fetch successful Sep 4 17:40:42.097443 coreos-metadata[910]: Sep 04 17:40:42.093 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Sep 4 17:40:42.105847 coreos-metadata[910]: Sep 04 17:40:42.105 INFO Fetch successful Sep 4 17:40:42.120593 coreos-metadata[910]: Sep 04 17:40:42.120 INFO wrote hostname ci-4054.1.0-a-b8703b3246 to /sysroot/etc/hostname Sep 4 17:40:42.122547 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 4 17:40:42.155377 initrd-setup-root[937]: cut: /sysroot/etc/passwd: No such file or directory Sep 4 17:40:42.207928 initrd-setup-root[944]: cut: /sysroot/etc/group: No such file or directory Sep 4 17:40:42.215051 initrd-setup-root[951]: cut: /sysroot/etc/shadow: No such file or directory Sep 4 17:40:42.221498 initrd-setup-root[958]: cut: /sysroot/etc/gshadow: No such file or directory Sep 4 17:40:43.214080 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 4 17:40:43.224970 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 4 17:40:43.232056 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 4 17:40:43.239042 kernel: BTRFS info (device sda6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:43.242364 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 4 17:40:43.264632 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 4 17:40:43.271835 ignition[1028]: INFO : Ignition 2.19.0 Sep 4 17:40:43.271835 ignition[1028]: INFO : Stage: mount Sep 4 17:40:43.279132 ignition[1028]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:43.279132 ignition[1028]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:43.279132 ignition[1028]: INFO : mount: mount passed Sep 4 17:40:43.279132 ignition[1028]: INFO : Ignition finished successfully Sep 4 17:40:43.274049 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 4 17:40:43.282992 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 4 17:40:43.300818 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:40:43.322903 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (1039) Sep 4 17:40:43.327900 kernel: BTRFS info (device sda6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:40:43.327936 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:40:43.332509 kernel: BTRFS info (device sda6): using free space tree Sep 4 17:40:43.343945 kernel: BTRFS info (device sda6): auto enabling async discard Sep 4 17:40:43.345489 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:40:43.368594 ignition[1056]: INFO : Ignition 2.19.0 Sep 4 17:40:43.368594 ignition[1056]: INFO : Stage: files Sep 4 17:40:43.373007 ignition[1056]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:43.373007 ignition[1056]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:43.373007 ignition[1056]: DEBUG : files: compiled without relabeling support, skipping Sep 4 17:40:43.373007 ignition[1056]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 4 17:40:43.373007 ignition[1056]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 4 17:40:43.437333 ignition[1056]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 4 17:40:43.442208 ignition[1056]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 4 17:40:43.442208 ignition[1056]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 4 17:40:43.437863 unknown[1056]: wrote ssh authorized keys file for user: core Sep 4 17:40:43.512492 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:40:43.519588 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 4 17:40:43.557630 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 4 17:40:43.608016 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:40:43.614257 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 4 17:40:43.619325 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 4 17:40:43.619325 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:40:43.630940 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:40:43.635508 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:40:43.640237 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:43.645184 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Sep 4 17:40:44.174387 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 4 17:40:44.466827 ignition[1056]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Sep 4 17:40:44.466827 ignition[1056]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 4 17:40:44.484015 ignition[1056]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:40:44.489202 ignition[1056]: INFO : files: files passed Sep 4 17:40:44.489202 ignition[1056]: INFO : Ignition finished successfully Sep 4 17:40:44.485823 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 4 17:40:44.531351 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 4 17:40:44.538278 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 4 17:40:44.541713 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 4 17:40:44.541796 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 4 17:40:44.573286 initrd-setup-root-after-ignition[1084]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.573286 initrd-setup-root-after-ignition[1084]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.581630 initrd-setup-root-after-ignition[1088]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:40:44.587341 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:40:44.587915 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 4 17:40:44.600125 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 4 17:40:44.625261 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 4 17:40:44.625372 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 4 17:40:44.631430 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 4 17:40:44.637405 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 4 17:40:44.640252 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 4 17:40:44.641058 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 4 17:40:44.661236 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:40:44.670329 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 4 17:40:44.682462 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:40:44.686302 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:40:44.695388 systemd[1]: Stopped target timers.target - Timer Units. Sep 4 17:40:44.700306 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 4 17:40:44.700478 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:40:44.707011 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 4 17:40:44.712263 systemd[1]: Stopped target basic.target - Basic System. Sep 4 17:40:44.719872 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 4 17:40:44.722964 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:40:44.731877 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 4 17:40:44.732100 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 4 17:40:44.732526 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:40:44.732985 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 4 17:40:44.733411 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 4 17:40:44.733846 systemd[1]: Stopped target swap.target - Swaps. Sep 4 17:40:44.734250 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 4 17:40:44.734383 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:40:44.735158 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:40:44.735616 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:40:44.736443 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 4 17:40:44.761581 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:40:44.764950 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 4 17:40:44.765088 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 4 17:40:44.771129 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 4 17:40:44.771275 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:40:44.777169 systemd[1]: ignition-files.service: Deactivated successfully. Sep 4 17:40:44.777313 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 4 17:40:44.782266 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Sep 4 17:40:44.794340 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Sep 4 17:40:44.835421 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 4 17:40:44.840065 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 4 17:40:44.845693 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 4 17:40:44.846071 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:40:44.858042 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 4 17:40:44.858192 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:40:44.866740 ignition[1108]: INFO : Ignition 2.19.0 Sep 4 17:40:44.866740 ignition[1108]: INFO : Stage: umount Sep 4 17:40:44.872775 ignition[1108]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:40:44.872775 ignition[1108]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Sep 4 17:40:44.872775 ignition[1108]: INFO : umount: umount passed Sep 4 17:40:44.872775 ignition[1108]: INFO : Ignition finished successfully Sep 4 17:40:44.869815 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 4 17:40:44.869915 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 4 17:40:44.876388 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 4 17:40:44.876495 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 4 17:40:44.881560 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 4 17:40:44.881609 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 4 17:40:44.892388 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 4 17:40:44.892439 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 4 17:40:44.901148 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 4 17:40:44.901200 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 4 17:40:44.906540 systemd[1]: Stopped target network.target - Network. Sep 4 17:40:44.911172 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 4 17:40:44.911229 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:40:44.914343 systemd[1]: Stopped target paths.target - Path Units. Sep 4 17:40:44.914437 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 4 17:40:44.919284 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:40:44.919370 systemd[1]: Stopped target slices.target - Slice Units. Sep 4 17:40:44.919809 systemd[1]: Stopped target sockets.target - Socket Units. Sep 4 17:40:44.920283 systemd[1]: iscsid.socket: Deactivated successfully. Sep 4 17:40:44.920325 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:40:44.920707 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 4 17:40:44.920739 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:40:44.921567 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 4 17:40:44.921607 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 4 17:40:44.922025 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 4 17:40:44.922059 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 4 17:40:44.922588 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 4 17:40:44.923065 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 4 17:40:44.950950 systemd-networkd[867]: eth0: DHCPv6 lease lost Sep 4 17:40:44.952665 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 4 17:40:44.952770 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 4 17:40:44.968548 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 4 17:40:44.969146 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 4 17:40:44.969270 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 4 17:40:44.974848 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 4 17:40:44.974943 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:40:44.993041 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 4 17:40:44.999191 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 4 17:40:44.999250 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:40:45.005133 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 17:40:45.005180 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:40:45.005275 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 4 17:40:45.005312 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 4 17:40:45.064599 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 4 17:40:45.064671 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:40:45.074053 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:40:45.091470 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 4 17:40:45.091628 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:40:45.098730 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 4 17:40:45.098769 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 4 17:40:45.104867 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 4 17:40:45.104927 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:40:45.121303 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 4 17:40:45.121368 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:40:45.133524 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 4 17:40:45.136045 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: Data path switched from VF: enP51082s1 Sep 4 17:40:45.133592 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 4 17:40:45.139690 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:40:45.139734 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:40:45.154094 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 4 17:40:45.160014 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 4 17:40:45.160074 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:40:45.163410 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 4 17:40:45.163456 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:40:45.179125 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 4 17:40:45.179178 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:40:45.187857 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:40:45.187928 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:45.196454 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 4 17:40:45.196558 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 4 17:40:45.204566 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 4 17:40:45.204677 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 4 17:40:45.444074 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 4 17:40:45.444204 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 4 17:40:45.452473 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 4 17:40:45.456947 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 4 17:40:45.457014 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 4 17:40:45.468100 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 4 17:40:45.475119 systemd[1]: Switching root. Sep 4 17:40:45.567469 systemd-journald[176]: Journal stopped Sep 4 17:40:51.862967 systemd-journald[176]: Received SIGTERM from PID 1 (systemd). Sep 4 17:40:51.863024 kernel: SELinux: policy capability network_peer_controls=1 Sep 4 17:40:51.863047 kernel: SELinux: policy capability open_perms=1 Sep 4 17:40:51.863067 kernel: SELinux: policy capability extended_socket_class=1 Sep 4 17:40:51.863085 kernel: SELinux: policy capability always_check_network=0 Sep 4 17:40:51.863105 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 4 17:40:51.863122 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 4 17:40:51.863147 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 4 17:40:51.863169 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 4 17:40:51.863188 kernel: audit: type=1403 audit(1725471647.134:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 4 17:40:51.863209 systemd[1]: Successfully loaded SELinux policy in 120.599ms. Sep 4 17:40:51.863229 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 11.169ms. Sep 4 17:40:51.863249 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:40:51.863270 systemd[1]: Detected virtualization microsoft. Sep 4 17:40:51.863294 systemd[1]: Detected architecture x86-64. Sep 4 17:40:51.863313 systemd[1]: Detected first boot. Sep 4 17:40:51.863336 systemd[1]: Hostname set to . Sep 4 17:40:51.863356 systemd[1]: Initializing machine ID from random generator. Sep 4 17:40:51.863376 zram_generator::config[1148]: No configuration found. Sep 4 17:40:51.863408 systemd[1]: Populated /etc with preset unit settings. Sep 4 17:40:51.863428 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 4 17:40:51.863453 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 4 17:40:51.863473 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 4 17:40:51.863491 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 4 17:40:51.863506 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 4 17:40:51.863521 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 4 17:40:51.863540 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 4 17:40:51.863556 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 4 17:40:51.863573 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 4 17:40:51.863591 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 4 17:40:51.863607 systemd[1]: Created slice user.slice - User and Session Slice. Sep 4 17:40:51.863625 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:40:51.863642 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:40:51.863660 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 4 17:40:51.863680 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 4 17:40:51.863697 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 4 17:40:51.863715 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:40:51.863732 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 4 17:40:51.863749 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:40:51.863766 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 4 17:40:51.863787 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 4 17:40:51.863805 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 4 17:40:51.863825 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 4 17:40:51.863843 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:40:51.863860 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:40:51.863878 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:40:51.863909 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:40:51.863925 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 4 17:40:51.863939 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 4 17:40:51.863958 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:40:51.863979 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:40:51.863991 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:40:51.864005 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 4 17:40:51.864017 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 4 17:40:51.864033 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 4 17:40:51.864046 systemd[1]: Mounting media.mount - External Media Directory... Sep 4 17:40:51.864060 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:51.864071 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 4 17:40:51.864083 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 4 17:40:51.864096 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 4 17:40:51.864107 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 4 17:40:51.864119 systemd[1]: Reached target machines.target - Containers. Sep 4 17:40:51.864135 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 4 17:40:51.864145 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:40:51.864156 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:40:51.864169 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 4 17:40:51.864180 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:40:51.864191 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:40:51.864205 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:40:51.864215 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 4 17:40:51.864228 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:40:51.864240 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 4 17:40:51.864253 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 4 17:40:51.864264 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 4 17:40:51.864276 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 4 17:40:51.864286 systemd[1]: Stopped systemd-fsck-usr.service. Sep 4 17:40:51.864299 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:40:51.864311 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:40:51.864324 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 4 17:40:51.864339 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 4 17:40:51.864350 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:40:51.864361 systemd[1]: verity-setup.service: Deactivated successfully. Sep 4 17:40:51.864372 systemd[1]: Stopped verity-setup.service. Sep 4 17:40:51.864386 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:51.864397 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 4 17:40:51.864431 systemd-journald[1243]: Collecting audit messages is disabled. Sep 4 17:40:51.864462 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 4 17:40:51.864475 systemd-journald[1243]: Journal started Sep 4 17:40:51.864499 systemd-journald[1243]: Runtime Journal (/run/log/journal/360e36820f7e4367aae97654bac8eacd) is 8.0M, max 158.8M, 150.8M free. Sep 4 17:40:51.154211 systemd[1]: Queued start job for default target multi-user.target. Sep 4 17:40:51.276027 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Sep 4 17:40:51.276436 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 4 17:40:51.871918 kernel: fuse: init (API version 7.39) Sep 4 17:40:51.871953 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:40:51.878564 systemd[1]: Mounted media.mount - External Media Directory. Sep 4 17:40:51.881505 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 4 17:40:51.884669 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 4 17:40:51.887806 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 4 17:40:51.890821 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 4 17:40:51.923842 kernel: loop: module loaded Sep 4 17:40:51.894422 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:40:51.899683 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 4 17:40:51.899850 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 4 17:40:51.903512 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:40:51.903680 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:40:51.907385 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:40:51.907580 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:40:51.911414 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 4 17:40:51.911604 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 4 17:40:51.916585 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:40:51.920237 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 17:40:51.925057 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 4 17:40:51.930830 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:40:51.931598 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:40:51.945448 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 4 17:40:51.957962 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 4 17:40:51.974961 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 4 17:40:51.978826 kernel: ACPI: bus type drm_connector registered Sep 4 17:40:51.979058 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 4 17:40:51.979102 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:40:51.983699 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 4 17:40:51.993054 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 4 17:40:52.005471 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 4 17:40:52.010683 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:40:52.033275 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 4 17:40:52.037385 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 4 17:40:52.040486 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:40:52.044053 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 4 17:40:52.048848 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:40:52.060133 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:40:52.066151 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 4 17:40:52.079815 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:40:52.086277 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:40:52.086502 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:40:52.090419 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:40:52.094341 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 4 17:40:52.098030 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 4 17:40:52.101461 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 4 17:40:52.105398 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 4 17:40:52.115633 systemd-journald[1243]: Time spent on flushing to /var/log/journal/360e36820f7e4367aae97654bac8eacd is 53.767ms for 963 entries. Sep 4 17:40:52.115633 systemd-journald[1243]: System Journal (/var/log/journal/360e36820f7e4367aae97654bac8eacd) is 8.0M, max 2.6G, 2.6G free. Sep 4 17:40:52.246121 systemd-journald[1243]: Received client request to flush runtime journal. Sep 4 17:40:52.246182 kernel: loop0: detected capacity change from 0 to 61888 Sep 4 17:40:52.113966 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 4 17:40:52.131035 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 4 17:40:52.137627 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 4 17:40:52.153986 udevadm[1293]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 4 17:40:52.207186 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:40:52.247588 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 4 17:40:52.359174 systemd-tmpfiles[1284]: ACLs are not supported, ignoring. Sep 4 17:40:52.359203 systemd-tmpfiles[1284]: ACLs are not supported, ignoring. Sep 4 17:40:52.366118 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:40:52.371616 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 4 17:40:52.372237 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 4 17:40:52.382060 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 4 17:40:52.480645 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 4 17:40:52.495008 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:40:52.514548 systemd-tmpfiles[1303]: ACLs are not supported, ignoring. Sep 4 17:40:52.514574 systemd-tmpfiles[1303]: ACLs are not supported, ignoring. Sep 4 17:40:52.518860 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:40:52.631909 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 4 17:40:52.680789 kernel: loop1: detected capacity change from 0 to 210664 Sep 4 17:40:52.742916 kernel: loop2: detected capacity change from 0 to 89336 Sep 4 17:40:53.149918 kernel: loop3: detected capacity change from 0 to 140728 Sep 4 17:40:53.579868 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 4 17:40:53.594328 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:40:53.615136 systemd-udevd[1311]: Using default interface naming scheme 'v255'. Sep 4 17:40:53.779920 kernel: loop4: detected capacity change from 0 to 61888 Sep 4 17:40:53.790911 kernel: loop5: detected capacity change from 0 to 210664 Sep 4 17:40:53.803919 kernel: loop6: detected capacity change from 0 to 89336 Sep 4 17:40:53.816947 kernel: loop7: detected capacity change from 0 to 140728 Sep 4 17:40:53.831188 (sd-merge)[1313]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Sep 4 17:40:53.831733 (sd-merge)[1313]: Merged extensions into '/usr'. Sep 4 17:40:53.835365 systemd[1]: Reloading requested from client PID 1282 ('systemd-sysext') (unit systemd-sysext.service)... Sep 4 17:40:53.835384 systemd[1]: Reloading... Sep 4 17:40:53.921914 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1324) Sep 4 17:40:53.979960 zram_generator::config[1358]: No configuration found. Sep 4 17:40:53.990958 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1324) Sep 4 17:40:54.077949 kernel: hv_vmbus: registering driver hv_balloon Sep 4 17:40:54.084086 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Sep 4 17:40:54.084175 kernel: mousedev: PS/2 mouse device common for all mice Sep 4 17:40:54.099429 kernel: hv_vmbus: registering driver hyperv_fb Sep 4 17:40:54.102979 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Sep 4 17:40:54.112787 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Sep 4 17:40:54.118290 kernel: Console: switching to colour dummy device 80x25 Sep 4 17:40:54.125027 kernel: Console: switching to colour frame buffer device 128x48 Sep 4 17:40:54.391623 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (1318) Sep 4 17:40:54.450214 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:40:54.532077 kernel: kvm_intel: Using Hyper-V Enlightened VMCS Sep 4 17:40:54.581719 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 4 17:40:54.583989 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Sep 4 17:40:54.587698 systemd[1]: Reloading finished in 751 ms. Sep 4 17:40:54.614620 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:40:54.619329 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 4 17:40:54.668660 systemd[1]: Starting ensure-sysext.service... Sep 4 17:40:54.677129 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 4 17:40:54.682297 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:40:54.688063 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 17:40:54.694742 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:40:54.703121 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 4 17:40:54.723220 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 4 17:40:54.727819 systemd[1]: Reloading requested from client PID 1469 ('systemctl') (unit ensure-sysext.service)... Sep 4 17:40:54.727836 systemd[1]: Reloading... Sep 4 17:40:54.734768 systemd-tmpfiles[1472]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 4 17:40:54.735307 systemd-tmpfiles[1472]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 4 17:40:54.737092 systemd-tmpfiles[1472]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 4 17:40:54.737549 systemd-tmpfiles[1472]: ACLs are not supported, ignoring. Sep 4 17:40:54.737642 systemd-tmpfiles[1472]: ACLs are not supported, ignoring. Sep 4 17:40:54.742530 systemd-tmpfiles[1472]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:40:54.742548 systemd-tmpfiles[1472]: Skipping /boot Sep 4 17:40:54.759573 systemd-tmpfiles[1472]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:40:54.759685 systemd-tmpfiles[1472]: Skipping /boot Sep 4 17:40:54.826912 zram_generator::config[1505]: No configuration found. Sep 4 17:40:54.831660 lvm[1476]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:40:54.958489 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:40:55.044716 systemd[1]: Reloading finished in 316 ms. Sep 4 17:40:55.075002 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 4 17:40:55.079070 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:40:55.082726 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:40:55.086205 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 4 17:40:55.097973 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:40:55.112519 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:40:55.158264 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 4 17:40:55.164026 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 4 17:40:55.173000 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 4 17:40:55.173843 lvm[1579]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:40:55.180222 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:40:55.185999 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 4 17:40:55.201194 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 4 17:40:55.212122 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:55.212425 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:40:55.218128 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:40:55.232233 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:40:55.246584 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:40:55.249398 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:40:55.249694 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:55.251332 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 4 17:40:55.255316 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:40:55.255491 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:40:55.259116 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:40:55.259290 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:40:55.263025 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:40:55.263183 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:40:55.272727 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 4 17:40:55.280328 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:55.280595 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:40:55.285150 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:40:55.289326 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:40:55.294813 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:40:55.301182 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:40:55.304193 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:40:55.304452 systemd[1]: Reached target time-set.target - System Time Set. Sep 4 17:40:55.307576 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:40:55.309017 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:40:55.309191 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:40:55.315328 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:40:55.315521 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:40:55.319230 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 4 17:40:55.325159 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:40:55.325934 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:40:55.333214 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:40:55.333396 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:40:55.343366 systemd[1]: Finished ensure-sysext.service. Sep 4 17:40:55.355404 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:40:55.355478 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:40:55.402947 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 4 17:40:55.408938 augenrules[1615]: No rules Sep 4 17:40:55.409781 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:40:55.438421 systemd-resolved[1581]: Positive Trust Anchors: Sep 4 17:40:55.438438 systemd-resolved[1581]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:40:55.438481 systemd-resolved[1581]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 17:40:55.461374 systemd-networkd[1471]: lo: Link UP Sep 4 17:40:55.461383 systemd-networkd[1471]: lo: Gained carrier Sep 4 17:40:55.463740 systemd-networkd[1471]: Enumeration completed Sep 4 17:40:55.463993 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:40:55.467760 systemd-networkd[1471]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:55.467769 systemd-networkd[1471]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:40:55.473836 systemd-resolved[1581]: Using system hostname 'ci-4054.1.0-a-b8703b3246'. Sep 4 17:40:55.476054 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 4 17:40:55.522899 kernel: mlx5_core c78a:00:02.0 enP51082s1: Link up Sep 4 17:40:55.547134 kernel: hv_netvsc 6045bdd0-ea60-6045-bdd0-ea606045bdd0 eth0: Data path switched to VF: enP51082s1 Sep 4 17:40:55.548786 systemd-networkd[1471]: enP51082s1: Link UP Sep 4 17:40:55.549213 systemd-networkd[1471]: eth0: Link UP Sep 4 17:40:55.549223 systemd-networkd[1471]: eth0: Gained carrier Sep 4 17:40:55.549254 systemd-networkd[1471]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:40:55.550863 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:40:55.554170 systemd[1]: Reached target network.target - Network. Sep 4 17:40:55.554255 systemd-networkd[1471]: enP51082s1: Gained carrier Sep 4 17:40:55.557105 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:40:55.597043 systemd-networkd[1471]: eth0: DHCPv4 address 10.200.4.34/24, gateway 10.200.4.1 acquired from 168.63.129.16 Sep 4 17:40:56.223043 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 4 17:40:56.227666 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 4 17:40:56.626329 systemd-networkd[1471]: eth0: Gained IPv6LL Sep 4 17:40:56.629961 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 4 17:40:56.634181 systemd[1]: Reached target network-online.target - Network is Online. Sep 4 17:40:56.690096 systemd-networkd[1471]: enP51082s1: Gained IPv6LL Sep 4 17:40:59.148742 ldconfig[1277]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 4 17:40:59.163746 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 4 17:40:59.172110 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 4 17:40:59.187013 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 4 17:40:59.190274 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:40:59.193325 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 4 17:40:59.196980 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 4 17:40:59.200592 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 4 17:40:59.203422 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 4 17:40:59.207194 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 4 17:40:59.210564 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 4 17:40:59.210601 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:40:59.213021 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:40:59.216539 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 4 17:40:59.220836 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 4 17:40:59.231741 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 4 17:40:59.235924 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 4 17:40:59.238762 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:40:59.241299 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:40:59.243750 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:40:59.243793 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:40:59.264991 systemd[1]: Starting chronyd.service - NTP client/server... Sep 4 17:40:59.272002 systemd[1]: Starting containerd.service - containerd container runtime... Sep 4 17:40:59.280061 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 4 17:40:59.290115 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 4 17:40:59.300012 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 4 17:40:59.308097 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 4 17:40:59.311083 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 4 17:40:59.315507 (chronyd)[1631]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Sep 4 17:40:59.316019 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:59.322605 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 4 17:40:59.323750 jq[1635]: false Sep 4 17:40:59.329586 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 4 17:40:59.334911 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 4 17:40:59.342062 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 4 17:40:59.344829 chronyd[1645]: chronyd version 4.5 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Sep 4 17:40:59.348151 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 4 17:40:59.367016 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 4 17:40:59.370491 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 4 17:40:59.371051 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 4 17:40:59.372056 systemd[1]: Starting update-engine.service - Update Engine... Sep 4 17:40:59.375997 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 4 17:40:59.381186 chronyd[1645]: Timezone right/UTC failed leap second check, ignoring Sep 4 17:40:59.381280 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 4 17:40:59.381400 chronyd[1645]: Loaded seccomp filter (level 2) Sep 4 17:40:59.381694 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 4 17:40:59.391345 systemd[1]: Started chronyd.service - NTP client/server. Sep 4 17:40:59.420859 extend-filesystems[1638]: Found loop4 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found loop5 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found loop6 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found loop7 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda1 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda2 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda3 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found usr Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda4 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda6 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda7 Sep 4 17:40:59.428415 extend-filesystems[1638]: Found sda9 Sep 4 17:40:59.428415 extend-filesystems[1638]: Checking size of /dev/sda9 Sep 4 17:40:59.444529 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 4 17:40:59.444763 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 4 17:40:59.462201 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 4 17:40:59.478386 systemd[1]: motdgen.service: Deactivated successfully. Sep 4 17:40:59.478601 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 4 17:40:59.511026 jq[1653]: true Sep 4 17:40:59.496275 (ntainerd)[1676]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 4 17:40:59.517690 dbus-daemon[1634]: [system] SELinux support is enabled Sep 4 17:40:59.528544 jq[1678]: true Sep 4 17:40:59.528699 extend-filesystems[1638]: Old size kept for /dev/sda9 Sep 4 17:40:59.528699 extend-filesystems[1638]: Found sr0 Sep 4 17:40:59.517863 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 4 17:40:59.526753 systemd-logind[1648]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 4 17:40:59.537969 systemd-logind[1648]: New seat seat0. Sep 4 17:40:59.571817 tar[1659]: linux-amd64/helm Sep 4 17:40:59.546360 systemd[1]: Started systemd-logind.service - User Login Management. Sep 4 17:40:59.551587 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 4 17:40:59.552606 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 4 17:40:59.567665 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 4 17:40:59.567723 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 4 17:40:59.572748 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 4 17:40:59.580631 dbus-daemon[1634]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 4 17:40:59.572771 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 4 17:40:59.607418 update_engine[1652]: I0904 17:40:59.606830 1652 main.cc:92] Flatcar Update Engine starting Sep 4 17:40:59.617075 systemd[1]: Started update-engine.service - Update Engine. Sep 4 17:40:59.623938 update_engine[1652]: I0904 17:40:59.623065 1652 update_check_scheduler.cc:74] Next update check in 8m16s Sep 4 17:40:59.628466 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 4 17:40:59.688950 coreos-metadata[1633]: Sep 04 17:40:59.686 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Sep 4 17:40:59.693227 coreos-metadata[1633]: Sep 04 17:40:59.692 INFO Fetch successful Sep 4 17:40:59.693227 coreos-metadata[1633]: Sep 04 17:40:59.692 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Sep 4 17:40:59.698908 coreos-metadata[1633]: Sep 04 17:40:59.698 INFO Fetch successful Sep 4 17:40:59.698908 coreos-metadata[1633]: Sep 04 17:40:59.698 INFO Fetching http://168.63.129.16/machine/5e47f4f6-38c0-4b3d-be68-802b3da0f51d/2d47a4a8%2D21eb%2D4cb7%2D8ec9%2Da4d92e82ba01.%5Fci%2D4054.1.0%2Da%2Db8703b3246?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Sep 4 17:40:59.702552 coreos-metadata[1633]: Sep 04 17:40:59.702 INFO Fetch successful Sep 4 17:40:59.702552 coreos-metadata[1633]: Sep 04 17:40:59.702 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Sep 4 17:40:59.720203 coreos-metadata[1633]: Sep 04 17:40:59.717 INFO Fetch successful Sep 4 17:40:59.733914 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (1701) Sep 4 17:40:59.796396 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 4 17:40:59.804675 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 4 17:40:59.807208 bash[1719]: Updated "/home/core/.ssh/authorized_keys" Sep 4 17:40:59.811173 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 4 17:40:59.819736 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 4 17:41:00.030341 locksmithd[1690]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 4 17:41:00.159059 sshd_keygen[1673]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 4 17:41:00.191132 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 4 17:41:00.207164 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 4 17:41:00.216083 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Sep 4 17:41:00.249868 systemd[1]: issuegen.service: Deactivated successfully. Sep 4 17:41:00.250148 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 4 17:41:00.261026 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Sep 4 17:41:00.275172 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 4 17:41:00.327239 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 4 17:41:00.341452 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 4 17:41:00.354526 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 4 17:41:00.358159 systemd[1]: Reached target getty.target - Login Prompts. Sep 4 17:41:00.481282 tar[1659]: linux-amd64/LICENSE Sep 4 17:41:00.481408 tar[1659]: linux-amd64/README.md Sep 4 17:41:00.493770 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 4 17:41:00.790815 containerd[1676]: time="2024-09-04T17:41:00.789879000Z" level=info msg="starting containerd" revision=8ccfc03e4e2b73c22899202ae09d0caf906d3863 version=v1.7.20 Sep 4 17:41:00.824111 containerd[1676]: time="2024-09-04T17:41:00.823902700Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.826790 containerd[1676]: time="2024-09-04T17:41:00.826749600Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.48-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:41:00.826790 containerd[1676]: time="2024-09-04T17:41:00.826779300Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 4 17:41:00.826929 containerd[1676]: time="2024-09-04T17:41:00.826798900Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 4 17:41:00.827323 containerd[1676]: time="2024-09-04T17:41:00.827015300Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 4 17:41:00.827323 containerd[1676]: time="2024-09-04T17:41:00.827059200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827323 containerd[1676]: time="2024-09-04T17:41:00.827155100Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827323 containerd[1676]: time="2024-09-04T17:41:00.827173900Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827502 containerd[1676]: time="2024-09-04T17:41:00.827378200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827502 containerd[1676]: time="2024-09-04T17:41:00.827402500Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827502 containerd[1676]: time="2024-09-04T17:41:00.827451200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827502 containerd[1676]: time="2024-09-04T17:41:00.827468400Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827639 containerd[1676]: time="2024-09-04T17:41:00.827574900Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.827996 containerd[1676]: time="2024-09-04T17:41:00.827824500Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:41:00.828062 containerd[1676]: time="2024-09-04T17:41:00.827994600Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:41:00.828062 containerd[1676]: time="2024-09-04T17:41:00.828015100Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 4 17:41:00.828143 containerd[1676]: time="2024-09-04T17:41:00.828121700Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 4 17:41:00.828229 containerd[1676]: time="2024-09-04T17:41:00.828174800Z" level=info msg="metadata content store policy set" policy=shared Sep 4 17:41:00.842667 containerd[1676]: time="2024-09-04T17:41:00.842629900Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 4 17:41:00.842742 containerd[1676]: time="2024-09-04T17:41:00.842691700Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 4 17:41:00.842742 containerd[1676]: time="2024-09-04T17:41:00.842716000Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 4 17:41:00.842742 containerd[1676]: time="2024-09-04T17:41:00.842736400Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 4 17:41:00.842845 containerd[1676]: time="2024-09-04T17:41:00.842757300Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.842980400Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843266500Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843395900Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843417200Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843435600Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843454100Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843471900Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843489900Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843508700Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843528600Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843546200Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843562900Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843578600Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 4 17:41:00.844445 containerd[1676]: time="2024-09-04T17:41:00.843603400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843622400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843640100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843658400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843676100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843702500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843726500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843746400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843764200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843784100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843810800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843829000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843845900Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843866800Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843908800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.844998 containerd[1676]: time="2024-09-04T17:41:00.843925800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.843941400Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844015500Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844047400Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844061700Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844077600Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844090900Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844104700Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844117400Z" level=info msg="NRI interface is disabled by configuration." Sep 4 17:41:00.845639 containerd[1676]: time="2024-09-04T17:41:00.844130500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.844500700Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.844586400Z" level=info msg="Connect containerd service" Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.844641200Z" level=info msg="using legacy CRI server" Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.844651100Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.844814500Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 4 17:41:00.845960 containerd[1676]: time="2024-09-04T17:41:00.845620500Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 17:41:00.847594 containerd[1676]: time="2024-09-04T17:41:00.847466500Z" level=info msg="Start subscribing containerd event" Sep 4 17:41:00.847594 containerd[1676]: time="2024-09-04T17:41:00.847545000Z" level=info msg="Start recovering state" Sep 4 17:41:00.848351 containerd[1676]: time="2024-09-04T17:41:00.847766600Z" level=info msg="Start event monitor" Sep 4 17:41:00.848351 containerd[1676]: time="2024-09-04T17:41:00.847797000Z" level=info msg="Start snapshots syncer" Sep 4 17:41:00.848351 containerd[1676]: time="2024-09-04T17:41:00.847810700Z" level=info msg="Start cni network conf syncer for default" Sep 4 17:41:00.848351 containerd[1676]: time="2024-09-04T17:41:00.847827500Z" level=info msg="Start streaming server" Sep 4 17:41:00.850839 containerd[1676]: time="2024-09-04T17:41:00.848578800Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 4 17:41:00.850839 containerd[1676]: time="2024-09-04T17:41:00.848657100Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 4 17:41:00.850839 containerd[1676]: time="2024-09-04T17:41:00.849279100Z" level=info msg="containerd successfully booted in 0.060779s" Sep 4 17:41:00.849469 systemd[1]: Started containerd.service - containerd container runtime. Sep 4 17:41:01.060263 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:01.065256 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 4 17:41:01.067133 (kubelet)[1795]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:01.069496 systemd[1]: Startup finished in 857ms (firmware) + 29.698s (loader) + 993ms (kernel) + 11.100s (initrd) + 14.054s (userspace) = 56.705s. Sep 4 17:41:01.740702 kubelet[1795]: E0904 17:41:01.740651 1795 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:01.743045 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:01.743242 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:01.761682 login[1781]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 4 17:41:01.763551 login[1782]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Sep 4 17:41:01.771602 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 4 17:41:01.779268 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 4 17:41:01.782660 systemd-logind[1648]: New session 2 of user core. Sep 4 17:41:01.789343 systemd-logind[1648]: New session 1 of user core. Sep 4 17:41:01.796317 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 4 17:41:01.813193 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 4 17:41:01.817558 (systemd)[1809]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:41:01.962610 systemd[1809]: Queued start job for default target default.target. Sep 4 17:41:01.967837 systemd[1809]: Created slice app.slice - User Application Slice. Sep 4 17:41:01.967875 systemd[1809]: Reached target paths.target - Paths. Sep 4 17:41:01.968086 systemd[1809]: Reached target timers.target - Timers. Sep 4 17:41:01.969429 systemd[1809]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 4 17:41:01.980715 systemd[1809]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 4 17:41:01.980845 systemd[1809]: Reached target sockets.target - Sockets. Sep 4 17:41:01.980864 systemd[1809]: Reached target basic.target - Basic System. Sep 4 17:41:01.980928 systemd[1809]: Reached target default.target - Main User Target. Sep 4 17:41:01.980967 systemd[1809]: Startup finished in 155ms. Sep 4 17:41:01.981299 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 4 17:41:01.993077 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 4 17:41:01.994141 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 4 17:41:02.315206 waagent[1773]: 2024-09-04T17:41:02.315035Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.315458Z INFO Daemon Daemon OS: flatcar 4054.1.0 Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.316408Z INFO Daemon Daemon Python: 3.11.9 Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.317542Z INFO Daemon Daemon Run daemon Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.318442Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4054.1.0' Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.319342Z INFO Daemon Daemon Using waagent for provisioning Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.319938Z INFO Daemon Daemon Activate resource disk Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.320259Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.324205Z INFO Daemon Daemon Found device: None Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.324938Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.325408Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.328197Z INFO Daemon Daemon Clean protocol and wireserver endpoint Sep 4 17:41:02.353202 waagent[1773]: 2024-09-04T17:41:02.329203Z INFO Daemon Daemon Running default provisioning handler Sep 4 17:41:02.356381 waagent[1773]: 2024-09-04T17:41:02.356306Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Sep 4 17:41:02.363144 waagent[1773]: 2024-09-04T17:41:02.363086Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Sep 4 17:41:02.368129 waagent[1773]: 2024-09-04T17:41:02.368035Z INFO Daemon Daemon cloud-init is enabled: False Sep 4 17:41:02.372544 waagent[1773]: 2024-09-04T17:41:02.368239Z INFO Daemon Daemon Copying ovf-env.xml Sep 4 17:41:02.455684 waagent[1773]: 2024-09-04T17:41:02.453147Z INFO Daemon Daemon Successfully mounted dvd Sep 4 17:41:02.487063 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Sep 4 17:41:02.488952 waagent[1773]: 2024-09-04T17:41:02.488812Z INFO Daemon Daemon Detect protocol endpoint Sep 4 17:41:02.491616 waagent[1773]: 2024-09-04T17:41:02.491469Z INFO Daemon Daemon Clean protocol and wireserver endpoint Sep 4 17:41:02.506203 waagent[1773]: 2024-09-04T17:41:02.491764Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Sep 4 17:41:02.506203 waagent[1773]: 2024-09-04T17:41:02.494311Z INFO Daemon Daemon Test for route to 168.63.129.16 Sep 4 17:41:02.506203 waagent[1773]: 2024-09-04T17:41:02.494925Z INFO Daemon Daemon Route to 168.63.129.16 exists Sep 4 17:41:02.506203 waagent[1773]: 2024-09-04T17:41:02.495261Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Sep 4 17:41:02.627538 waagent[1773]: 2024-09-04T17:41:02.627384Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Sep 4 17:41:02.631118 waagent[1773]: 2024-09-04T17:41:02.631076Z INFO Daemon Daemon Wire protocol version:2012-11-30 Sep 4 17:41:02.633932 waagent[1773]: 2024-09-04T17:41:02.633872Z INFO Daemon Daemon Server preferred version:2015-04-05 Sep 4 17:41:02.774953 waagent[1773]: 2024-09-04T17:41:02.774830Z INFO Daemon Daemon Initializing goal state during protocol detection Sep 4 17:41:02.784411 waagent[1773]: 2024-09-04T17:41:02.775207Z INFO Daemon Daemon Forcing an update of the goal state. Sep 4 17:41:02.784411 waagent[1773]: 2024-09-04T17:41:02.779329Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Sep 4 17:41:02.796403 waagent[1773]: 2024-09-04T17:41:02.796345Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.159 Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.796964Z INFO Daemon Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.797564Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: a747f3b9-de8e-4062-9c47-abe28abd62f7 eTag: 12371950208546475057 source: Fabric] Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.798761Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.799993Z INFO Daemon Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.800514Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Sep 4 17:41:02.807809 waagent[1773]: 2024-09-04T17:41:02.804776Z INFO Daemon Daemon Downloading artifacts profile blob Sep 4 17:41:02.890339 waagent[1773]: 2024-09-04T17:41:02.890200Z INFO Daemon Downloaded certificate {'thumbprint': 'D1875B999C5E97A42C6BE50583B8465C03864203', 'hasPrivateKey': True} Sep 4 17:41:02.895796 waagent[1773]: 2024-09-04T17:41:02.895732Z INFO Daemon Downloaded certificate {'thumbprint': 'A5687415CE9DA2CD5F6F3463A06613BE6404DDF8', 'hasPrivateKey': False} Sep 4 17:41:02.901033 waagent[1773]: 2024-09-04T17:41:02.900977Z INFO Daemon Fetch goal state completed Sep 4 17:41:02.909555 waagent[1773]: 2024-09-04T17:41:02.909504Z INFO Daemon Daemon Starting provisioning Sep 4 17:41:02.916449 waagent[1773]: 2024-09-04T17:41:02.909712Z INFO Daemon Daemon Handle ovf-env.xml. Sep 4 17:41:02.916449 waagent[1773]: 2024-09-04T17:41:02.910737Z INFO Daemon Daemon Set hostname [ci-4054.1.0-a-b8703b3246] Sep 4 17:41:02.934945 waagent[1773]: 2024-09-04T17:41:02.934845Z INFO Daemon Daemon Publish hostname [ci-4054.1.0-a-b8703b3246] Sep 4 17:41:02.943112 waagent[1773]: 2024-09-04T17:41:02.935352Z INFO Daemon Daemon Examine /proc/net/route for primary interface Sep 4 17:41:02.943112 waagent[1773]: 2024-09-04T17:41:02.935836Z INFO Daemon Daemon Primary interface is [eth0] Sep 4 17:41:02.961749 systemd-networkd[1471]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:41:02.961761 systemd-networkd[1471]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:41:02.961814 systemd-networkd[1471]: eth0: DHCP lease lost Sep 4 17:41:02.963326 waagent[1773]: 2024-09-04T17:41:02.963245Z INFO Daemon Daemon Create user account if not exists Sep 4 17:41:02.966581 waagent[1773]: 2024-09-04T17:41:02.966456Z INFO Daemon Daemon User core already exists, skip useradd Sep 4 17:41:02.981317 waagent[1773]: 2024-09-04T17:41:02.966659Z INFO Daemon Daemon Configure sudoer Sep 4 17:41:02.981317 waagent[1773]: 2024-09-04T17:41:02.968034Z INFO Daemon Daemon Configure sshd Sep 4 17:41:02.981317 waagent[1773]: 2024-09-04T17:41:02.968916Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Sep 4 17:41:02.981317 waagent[1773]: 2024-09-04T17:41:02.969195Z INFO Daemon Daemon Deploy ssh public key. Sep 4 17:41:02.981983 systemd-networkd[1471]: eth0: DHCPv6 lease lost Sep 4 17:41:03.008960 systemd-networkd[1471]: eth0: DHCPv4 address 10.200.4.34/24, gateway 10.200.4.1 acquired from 168.63.129.16 Sep 4 17:41:04.259544 waagent[1773]: 2024-09-04T17:41:04.259450Z INFO Daemon Daemon Provisioning complete Sep 4 17:41:04.272004 waagent[1773]: 2024-09-04T17:41:04.271950Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Sep 4 17:41:04.279047 waagent[1773]: 2024-09-04T17:41:04.272237Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Sep 4 17:41:04.279047 waagent[1773]: 2024-09-04T17:41:04.273124Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Sep 4 17:41:04.399463 waagent[1865]: 2024-09-04T17:41:04.399368Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Sep 4 17:41:04.399923 waagent[1865]: 2024-09-04T17:41:04.399530Z INFO ExtHandler ExtHandler OS: flatcar 4054.1.0 Sep 4 17:41:04.399923 waagent[1865]: 2024-09-04T17:41:04.399613Z INFO ExtHandler ExtHandler Python: 3.11.9 Sep 4 17:41:04.421629 waagent[1865]: 2024-09-04T17:41:04.421546Z INFO ExtHandler ExtHandler Distro: flatcar-4054.1.0; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.9; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Sep 4 17:41:04.421830 waagent[1865]: 2024-09-04T17:41:04.421784Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Sep 4 17:41:04.421948 waagent[1865]: 2024-09-04T17:41:04.421877Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Sep 4 17:41:04.429233 waagent[1865]: 2024-09-04T17:41:04.429165Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Sep 4 17:41:04.433969 waagent[1865]: 2024-09-04T17:41:04.433922Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.159 Sep 4 17:41:04.434414 waagent[1865]: 2024-09-04T17:41:04.434356Z INFO ExtHandler Sep 4 17:41:04.434516 waagent[1865]: 2024-09-04T17:41:04.434452Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 2cc87f22-4bd4-469f-bbdf-2e5ece7a8388 eTag: 12371950208546475057 source: Fabric] Sep 4 17:41:04.434814 waagent[1865]: 2024-09-04T17:41:04.434766Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Sep 4 17:41:04.435413 waagent[1865]: 2024-09-04T17:41:04.435350Z INFO ExtHandler Sep 4 17:41:04.435483 waagent[1865]: 2024-09-04T17:41:04.435438Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Sep 4 17:41:04.438660 waagent[1865]: 2024-09-04T17:41:04.438608Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Sep 4 17:41:04.535970 waagent[1865]: 2024-09-04T17:41:04.534243Z INFO ExtHandler Downloaded certificate {'thumbprint': 'D1875B999C5E97A42C6BE50583B8465C03864203', 'hasPrivateKey': True} Sep 4 17:41:04.535970 waagent[1865]: 2024-09-04T17:41:04.534834Z INFO ExtHandler Downloaded certificate {'thumbprint': 'A5687415CE9DA2CD5F6F3463A06613BE6404DDF8', 'hasPrivateKey': False} Sep 4 17:41:04.535970 waagent[1865]: 2024-09-04T17:41:04.535407Z INFO ExtHandler Fetch goal state completed Sep 4 17:41:04.549903 waagent[1865]: 2024-09-04T17:41:04.549807Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 1865 Sep 4 17:41:04.550050 waagent[1865]: 2024-09-04T17:41:04.550014Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Sep 4 17:41:04.551643 waagent[1865]: 2024-09-04T17:41:04.551579Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4054.1.0', '', 'Flatcar Container Linux by Kinvolk'] Sep 4 17:41:04.552048 waagent[1865]: 2024-09-04T17:41:04.551998Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Sep 4 17:41:05.198870 waagent[1865]: 2024-09-04T17:41:05.198818Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Sep 4 17:41:05.199137 waagent[1865]: 2024-09-04T17:41:05.199081Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Sep 4 17:41:05.205686 waagent[1865]: 2024-09-04T17:41:05.205640Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Sep 4 17:41:05.212412 systemd[1]: Reloading requested from client PID 1880 ('systemctl') (unit waagent.service)... Sep 4 17:41:05.212428 systemd[1]: Reloading... Sep 4 17:41:05.297911 zram_generator::config[1908]: No configuration found. Sep 4 17:41:05.468895 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:41:05.544875 systemd[1]: Reloading finished in 331 ms. Sep 4 17:41:05.569944 waagent[1865]: 2024-09-04T17:41:05.569476Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Sep 4 17:41:05.578500 systemd[1]: Reloading requested from client PID 1968 ('systemctl') (unit waagent.service)... Sep 4 17:41:05.578514 systemd[1]: Reloading... Sep 4 17:41:05.658918 zram_generator::config[1999]: No configuration found. Sep 4 17:41:05.781721 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:41:05.856756 systemd[1]: Reloading finished in 277 ms. Sep 4 17:41:05.883583 waagent[1865]: 2024-09-04T17:41:05.883482Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Sep 4 17:41:05.884917 waagent[1865]: 2024-09-04T17:41:05.883698Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Sep 4 17:41:06.209310 waagent[1865]: 2024-09-04T17:41:06.209144Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Sep 4 17:41:06.209951 waagent[1865]: 2024-09-04T17:41:06.209851Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Sep 4 17:41:06.210860 waagent[1865]: 2024-09-04T17:41:06.210782Z INFO ExtHandler ExtHandler Starting env monitor service. Sep 4 17:41:06.211446 waagent[1865]: 2024-09-04T17:41:06.211371Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Sep 4 17:41:06.211594 waagent[1865]: 2024-09-04T17:41:06.211526Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Sep 4 17:41:06.211752 waagent[1865]: 2024-09-04T17:41:06.211693Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Sep 4 17:41:06.211961 waagent[1865]: 2024-09-04T17:41:06.211846Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Sep 4 17:41:06.212149 waagent[1865]: 2024-09-04T17:41:06.212087Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Sep 4 17:41:06.212490 waagent[1865]: 2024-09-04T17:41:06.212428Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Sep 4 17:41:06.212796 waagent[1865]: 2024-09-04T17:41:06.212740Z INFO EnvHandler ExtHandler Configure routes Sep 4 17:41:06.213023 waagent[1865]: 2024-09-04T17:41:06.212981Z INFO EnvHandler ExtHandler Gateway:None Sep 4 17:41:06.213119 waagent[1865]: 2024-09-04T17:41:06.213046Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Sep 4 17:41:06.213310 waagent[1865]: 2024-09-04T17:41:06.213263Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Sep 4 17:41:06.213487 waagent[1865]: 2024-09-04T17:41:06.213445Z INFO EnvHandler ExtHandler Routes:None Sep 4 17:41:06.213926 waagent[1865]: 2024-09-04T17:41:06.213857Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Sep 4 17:41:06.214054 waagent[1865]: 2024-09-04T17:41:06.213980Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Sep 4 17:41:06.214054 waagent[1865]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Sep 4 17:41:06.214054 waagent[1865]: eth0 00000000 0104C80A 0003 0 0 1024 00000000 0 0 0 Sep 4 17:41:06.214054 waagent[1865]: eth0 0004C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Sep 4 17:41:06.214054 waagent[1865]: eth0 0104C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Sep 4 17:41:06.214054 waagent[1865]: eth0 10813FA8 0104C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Sep 4 17:41:06.214054 waagent[1865]: eth0 FEA9FEA9 0104C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Sep 4 17:41:06.216920 waagent[1865]: 2024-09-04T17:41:06.214712Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Sep 4 17:41:06.216920 waagent[1865]: 2024-09-04T17:41:06.215310Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Sep 4 17:41:06.225910 waagent[1865]: 2024-09-04T17:41:06.224258Z INFO ExtHandler ExtHandler Sep 4 17:41:06.225910 waagent[1865]: 2024-09-04T17:41:06.224369Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: e5836cdb-2b45-4e66-97cf-af6a9f231bc1 correlation c24aba3f-e88c-495b-bdd2-86e294eaf9a0 created: 2024-09-04T17:39:54.485840Z] Sep 4 17:41:06.225910 waagent[1865]: 2024-09-04T17:41:06.224838Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Sep 4 17:41:06.225910 waagent[1865]: 2024-09-04T17:41:06.225733Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 1 ms] Sep 4 17:41:06.258843 waagent[1865]: 2024-09-04T17:41:06.258694Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 4F2C9283-BCA4-46AB-889B-55256E5123A0;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Sep 4 17:41:06.280847 waagent[1865]: 2024-09-04T17:41:06.280779Z INFO MonitorHandler ExtHandler Network interfaces: Sep 4 17:41:06.280847 waagent[1865]: Executing ['ip', '-a', '-o', 'link']: Sep 4 17:41:06.280847 waagent[1865]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Sep 4 17:41:06.280847 waagent[1865]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 60:45:bd:d0:ea:60 brd ff:ff:ff:ff:ff:ff Sep 4 17:41:06.280847 waagent[1865]: 3: enP51082s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 60:45:bd:d0:ea:60 brd ff:ff:ff:ff:ff:ff\ altname enP51082p0s2 Sep 4 17:41:06.280847 waagent[1865]: Executing ['ip', '-4', '-a', '-o', 'address']: Sep 4 17:41:06.280847 waagent[1865]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Sep 4 17:41:06.280847 waagent[1865]: 2: eth0 inet 10.200.4.34/24 metric 1024 brd 10.200.4.255 scope global eth0\ valid_lft forever preferred_lft forever Sep 4 17:41:06.280847 waagent[1865]: Executing ['ip', '-6', '-a', '-o', 'address']: Sep 4 17:41:06.280847 waagent[1865]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Sep 4 17:41:06.280847 waagent[1865]: 2: eth0 inet6 fe80::6245:bdff:fed0:ea60/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Sep 4 17:41:06.280847 waagent[1865]: 3: enP51082s1 inet6 fe80::6245:bdff:fed0:ea60/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Sep 4 17:41:06.309540 waagent[1865]: 2024-09-04T17:41:06.309489Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Sep 4 17:41:06.309540 waagent[1865]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.309540 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.309540 waagent[1865]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.309540 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.309540 waagent[1865]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.309540 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.309540 waagent[1865]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Sep 4 17:41:06.309540 waagent[1865]: 5 457 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Sep 4 17:41:06.309540 waagent[1865]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Sep 4 17:41:06.312859 waagent[1865]: 2024-09-04T17:41:06.312803Z INFO EnvHandler ExtHandler Current Firewall rules: Sep 4 17:41:06.312859 waagent[1865]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.312859 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.312859 waagent[1865]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.312859 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.312859 waagent[1865]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Sep 4 17:41:06.312859 waagent[1865]: pkts bytes target prot opt in out source destination Sep 4 17:41:06.312859 waagent[1865]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Sep 4 17:41:06.312859 waagent[1865]: 10 1102 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Sep 4 17:41:06.312859 waagent[1865]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Sep 4 17:41:06.313267 waagent[1865]: 2024-09-04T17:41:06.313127Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Sep 4 17:41:11.978406 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 4 17:41:11.991115 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:41:12.085631 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:12.090181 (kubelet)[2095]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:12.670064 kubelet[2095]: E0904 17:41:12.670006 2095 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:12.674067 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:12.674268 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:22.728455 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 4 17:41:22.734114 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:41:22.826984 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:22.831603 (kubelet)[2111]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:23.183765 chronyd[1645]: Selected source PHC0 Sep 4 17:41:23.444295 kubelet[2111]: E0904 17:41:23.444144 2111 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:23.446986 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:23.447179 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:31.793955 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 4 17:41:31.802173 systemd[1]: Started sshd@0-10.200.4.34:22-10.200.16.10:46082.service - OpenSSH per-connection server daemon (10.200.16.10:46082). Sep 4 17:41:32.439386 sshd[2120]: Accepted publickey for core from 10.200.16.10 port 46082 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:32.441220 sshd[2120]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:32.446774 systemd-logind[1648]: New session 3 of user core. Sep 4 17:41:32.458046 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 4 17:41:32.965818 systemd[1]: Started sshd@1-10.200.4.34:22-10.200.16.10:46096.service - OpenSSH per-connection server daemon (10.200.16.10:46096). Sep 4 17:41:33.449333 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Sep 4 17:41:33.456116 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:41:33.548483 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:33.553153 (kubelet)[2135]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:33.554551 sshd[2125]: Accepted publickey for core from 10.200.16.10 port 46096 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:33.556872 sshd[2125]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:33.563096 systemd-logind[1648]: New session 4 of user core. Sep 4 17:41:33.568279 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 4 17:41:33.983143 sshd[2125]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:33.986385 systemd[1]: sshd@1-10.200.4.34:22-10.200.16.10:46096.service: Deactivated successfully. Sep 4 17:41:33.988316 systemd[1]: session-4.scope: Deactivated successfully. Sep 4 17:41:33.989761 systemd-logind[1648]: Session 4 logged out. Waiting for processes to exit. Sep 4 17:41:33.990700 systemd-logind[1648]: Removed session 4. Sep 4 17:41:34.092187 systemd[1]: Started sshd@2-10.200.4.34:22-10.200.16.10:46110.service - OpenSSH per-connection server daemon (10.200.16.10:46110). Sep 4 17:41:34.132233 kubelet[2135]: E0904 17:41:34.132129 2135 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:34.134935 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:34.135193 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:34.670972 sshd[2147]: Accepted publickey for core from 10.200.16.10 port 46110 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:34.672719 sshd[2147]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:34.678305 systemd-logind[1648]: New session 5 of user core. Sep 4 17:41:34.688040 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 4 17:41:35.083188 sshd[2147]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:35.087682 systemd[1]: sshd@2-10.200.4.34:22-10.200.16.10:46110.service: Deactivated successfully. Sep 4 17:41:35.089538 systemd[1]: session-5.scope: Deactivated successfully. Sep 4 17:41:35.090292 systemd-logind[1648]: Session 5 logged out. Waiting for processes to exit. Sep 4 17:41:35.091266 systemd-logind[1648]: Removed session 5. Sep 4 17:41:35.195524 systemd[1]: Started sshd@3-10.200.4.34:22-10.200.16.10:46118.service - OpenSSH per-connection server daemon (10.200.16.10:46118). Sep 4 17:41:35.777091 sshd[2155]: Accepted publickey for core from 10.200.16.10 port 46118 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:35.778819 sshd[2155]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:35.783791 systemd-logind[1648]: New session 6 of user core. Sep 4 17:41:35.793055 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 4 17:41:36.248831 sshd[2155]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:36.253046 systemd[1]: sshd@3-10.200.4.34:22-10.200.16.10:46118.service: Deactivated successfully. Sep 4 17:41:36.254812 systemd[1]: session-6.scope: Deactivated successfully. Sep 4 17:41:36.255515 systemd-logind[1648]: Session 6 logged out. Waiting for processes to exit. Sep 4 17:41:36.256410 systemd-logind[1648]: Removed session 6. Sep 4 17:41:36.357195 systemd[1]: Started sshd@4-10.200.4.34:22-10.200.16.10:46128.service - OpenSSH per-connection server daemon (10.200.16.10:46128). Sep 4 17:41:36.936506 sshd[2162]: Accepted publickey for core from 10.200.16.10 port 46128 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:36.938259 sshd[2162]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:36.943801 systemd-logind[1648]: New session 7 of user core. Sep 4 17:41:36.953049 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 4 17:41:37.547830 sudo[2165]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 4 17:41:37.548222 sudo[2165]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:41:37.559233 sudo[2165]: pam_unix(sudo:session): session closed for user root Sep 4 17:41:37.652585 sshd[2162]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:37.656385 systemd[1]: sshd@4-10.200.4.34:22-10.200.16.10:46128.service: Deactivated successfully. Sep 4 17:41:37.658847 systemd[1]: session-7.scope: Deactivated successfully. Sep 4 17:41:37.660550 systemd-logind[1648]: Session 7 logged out. Waiting for processes to exit. Sep 4 17:41:37.661548 systemd-logind[1648]: Removed session 7. Sep 4 17:41:37.758131 systemd[1]: Started sshd@5-10.200.4.34:22-10.200.16.10:46130.service - OpenSSH per-connection server daemon (10.200.16.10:46130). Sep 4 17:41:38.348359 sshd[2173]: Accepted publickey for core from 10.200.16.10 port 46130 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:38.349857 sshd[2173]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:38.354342 systemd-logind[1648]: New session 8 of user core. Sep 4 17:41:38.361269 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 4 17:41:38.676319 sudo[2177]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 4 17:41:38.676685 sudo[2177]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:41:38.699283 sudo[2177]: pam_unix(sudo:session): session closed for user root Sep 4 17:41:38.705534 sudo[2176]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 4 17:41:38.705910 sudo[2176]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:41:38.724243 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 4 17:41:38.725840 auditctl[2180]: No rules Sep 4 17:41:38.726280 systemd[1]: audit-rules.service: Deactivated successfully. Sep 4 17:41:38.726487 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 4 17:41:38.729031 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:41:38.754813 augenrules[2198]: No rules Sep 4 17:41:38.756191 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:41:38.757731 sudo[2176]: pam_unix(sudo:session): session closed for user root Sep 4 17:41:38.859564 sshd[2173]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:38.863209 systemd[1]: sshd@5-10.200.4.34:22-10.200.16.10:46130.service: Deactivated successfully. Sep 4 17:41:38.865480 systemd[1]: session-8.scope: Deactivated successfully. Sep 4 17:41:38.867282 systemd-logind[1648]: Session 8 logged out. Waiting for processes to exit. Sep 4 17:41:38.868520 systemd-logind[1648]: Removed session 8. Sep 4 17:41:38.964067 systemd[1]: Started sshd@6-10.200.4.34:22-10.200.16.10:51878.service - OpenSSH per-connection server daemon (10.200.16.10:51878). Sep 4 17:41:39.553995 sshd[2206]: Accepted publickey for core from 10.200.16.10 port 51878 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:41:39.555496 sshd[2206]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:39.560109 systemd-logind[1648]: New session 9 of user core. Sep 4 17:41:39.570044 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 4 17:41:39.879754 sudo[2209]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 4 17:41:39.880219 sudo[2209]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:41:40.227227 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 4 17:41:40.227700 (dockerd)[2218]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 4 17:41:41.661659 dockerd[2218]: time="2024-09-04T17:41:41.661584701Z" level=info msg="Starting up" Sep 4 17:41:42.077156 dockerd[2218]: time="2024-09-04T17:41:42.077109843Z" level=info msg="Loading containers: start." Sep 4 17:41:42.166902 kernel: hv_balloon: Max. dynamic memory size: 8192 MB Sep 4 17:41:42.288050 kernel: Initializing XFRM netlink socket Sep 4 17:41:42.416986 systemd-networkd[1471]: docker0: Link UP Sep 4 17:41:42.447149 dockerd[2218]: time="2024-09-04T17:41:42.447102947Z" level=info msg="Loading containers: done." Sep 4 17:41:42.481705 dockerd[2218]: time="2024-09-04T17:41:42.481654783Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 4 17:41:42.482007 dockerd[2218]: time="2024-09-04T17:41:42.481811686Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 4 17:41:42.482007 dockerd[2218]: time="2024-09-04T17:41:42.481987689Z" level=info msg="Daemon has completed initialization" Sep 4 17:41:42.533641 dockerd[2218]: time="2024-09-04T17:41:42.533580238Z" level=info msg="API listen on /run/docker.sock" Sep 4 17:41:42.533786 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 4 17:41:44.228373 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Sep 4 17:41:44.240178 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:41:44.497564 containerd[1676]: time="2024-09-04T17:41:44.497242750Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.4\"" Sep 4 17:41:44.937320 update_engine[1652]: I0904 17:41:44.924042 1652 update_attempter.cc:509] Updating boot flags... Sep 4 17:41:44.947749 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:44.953516 (kubelet)[2369]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:45.005149 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (2384) Sep 4 17:41:45.029973 kubelet[2369]: E0904 17:41:45.028339 2369 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:45.032168 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:45.032352 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:45.967318 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1453459710.mount: Deactivated successfully. Sep 4 17:41:48.166514 containerd[1676]: time="2024-09-04T17:41:48.166452467Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:48.170266 containerd[1676]: time="2024-09-04T17:41:48.170199426Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.4: active requests=0, bytes read=32772424" Sep 4 17:41:48.174622 containerd[1676]: time="2024-09-04T17:41:48.174508595Z" level=info msg="ImageCreate event name:\"sha256:8a97b1fb3e2ebd03bf97ce8ae894b3dc8a68ab1f4ecfd0a284921c45c56f5aa4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:48.180001 containerd[1676]: time="2024-09-04T17:41:48.179962981Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:7b0c4a959aaee5660e1234452dc3123310231b9f92d29ebd175c86dc9f797ee7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:48.181222 containerd[1676]: time="2024-09-04T17:41:48.181008198Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.4\" with image id \"sha256:8a97b1fb3e2ebd03bf97ce8ae894b3dc8a68ab1f4ecfd0a284921c45c56f5aa4\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:7b0c4a959aaee5660e1234452dc3123310231b9f92d29ebd175c86dc9f797ee7\", size \"32769216\" in 3.683724346s" Sep 4 17:41:48.181222 containerd[1676]: time="2024-09-04T17:41:48.181055498Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.4\" returns image reference \"sha256:8a97b1fb3e2ebd03bf97ce8ae894b3dc8a68ab1f4ecfd0a284921c45c56f5aa4\"" Sep 4 17:41:48.203850 containerd[1676]: time="2024-09-04T17:41:48.203808659Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.4\"" Sep 4 17:41:50.397766 containerd[1676]: time="2024-09-04T17:41:50.397708413Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:50.401221 containerd[1676]: time="2024-09-04T17:41:50.401156068Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.4: active requests=0, bytes read=29594073" Sep 4 17:41:50.405175 containerd[1676]: time="2024-09-04T17:41:50.405121131Z" level=info msg="ImageCreate event name:\"sha256:8398ad49a121d58ecf8a36e8371c0928fdf75eb0a83d28232ab2b39b1c6a9050\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:50.411348 containerd[1676]: time="2024-09-04T17:41:50.411313529Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:992cccbf652fa951c1a3d41b0c1033ae0bf64f33da03d50395282c551900af9e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:50.413092 containerd[1676]: time="2024-09-04T17:41:50.413058656Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.4\" with image id \"sha256:8398ad49a121d58ecf8a36e8371c0928fdf75eb0a83d28232ab2b39b1c6a9050\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:992cccbf652fa951c1a3d41b0c1033ae0bf64f33da03d50395282c551900af9e\", size \"31144011\" in 2.209213697s" Sep 4 17:41:50.413092 containerd[1676]: time="2024-09-04T17:41:50.413090857Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.4\" returns image reference \"sha256:8398ad49a121d58ecf8a36e8371c0928fdf75eb0a83d28232ab2b39b1c6a9050\"" Sep 4 17:41:50.435932 containerd[1676]: time="2024-09-04T17:41:50.435895918Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.4\"" Sep 4 17:41:52.001455 containerd[1676]: time="2024-09-04T17:41:52.001394518Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:52.003383 containerd[1676]: time="2024-09-04T17:41:52.003330348Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.4: active requests=0, bytes read=17780241" Sep 4 17:41:52.008172 containerd[1676]: time="2024-09-04T17:41:52.008103624Z" level=info msg="ImageCreate event name:\"sha256:4939f82ab9ab456e782c06ed37b245127c8a9ac29a72982346a7160f18107833\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:52.014177 containerd[1676]: time="2024-09-04T17:41:52.014107319Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:37eaeee5bca8da34ad3d36e37586dd29f5edb1e2927e7644dfb113e70062bda8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:52.015905 containerd[1676]: time="2024-09-04T17:41:52.015081335Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.4\" with image id \"sha256:4939f82ab9ab456e782c06ed37b245127c8a9ac29a72982346a7160f18107833\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:37eaeee5bca8da34ad3d36e37586dd29f5edb1e2927e7644dfb113e70062bda8\", size \"19330197\" in 1.579146515s" Sep 4 17:41:52.015905 containerd[1676]: time="2024-09-04T17:41:52.015120335Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.4\" returns image reference \"sha256:4939f82ab9ab456e782c06ed37b245127c8a9ac29a72982346a7160f18107833\"" Sep 4 17:41:52.038786 containerd[1676]: time="2024-09-04T17:41:52.038745009Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.4\"" Sep 4 17:41:53.327655 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1004765685.mount: Deactivated successfully. Sep 4 17:41:53.887709 containerd[1676]: time="2024-09-04T17:41:53.887649099Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:53.889519 containerd[1676]: time="2024-09-04T17:41:53.889451227Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.4: active requests=0, bytes read=29037169" Sep 4 17:41:53.892855 containerd[1676]: time="2024-09-04T17:41:53.892802880Z" level=info msg="ImageCreate event name:\"sha256:568d5ba88d944bcd67415d8c358fce615824410f3a43bab2b353336bc3795a10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:53.897015 containerd[1676]: time="2024-09-04T17:41:53.896952046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:33ee1df1ba70e41bf9506d54bb5e64ef5f3ba9fc1b3021aaa4468606a7802acc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:53.897765 containerd[1676]: time="2024-09-04T17:41:53.897571756Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.4\" with image id \"sha256:568d5ba88d944bcd67415d8c358fce615824410f3a43bab2b353336bc3795a10\", repo tag \"registry.k8s.io/kube-proxy:v1.30.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:33ee1df1ba70e41bf9506d54bb5e64ef5f3ba9fc1b3021aaa4468606a7802acc\", size \"29036180\" in 1.858777646s" Sep 4 17:41:53.897765 containerd[1676]: time="2024-09-04T17:41:53.897610056Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.4\" returns image reference \"sha256:568d5ba88d944bcd67415d8c358fce615824410f3a43bab2b353336bc3795a10\"" Sep 4 17:41:53.921363 containerd[1676]: time="2024-09-04T17:41:53.921162830Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Sep 4 17:41:54.645105 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2358256188.mount: Deactivated successfully. Sep 4 17:41:55.228218 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Sep 4 17:41:55.238487 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:41:55.331310 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:41:55.335814 (kubelet)[2532]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:41:55.948310 kubelet[2532]: E0904 17:41:55.948232 2532 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:41:55.950587 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:41:55.950871 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:41:56.478557 containerd[1676]: time="2024-09-04T17:41:56.478435287Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:56.482078 containerd[1676]: time="2024-09-04T17:41:56.481949447Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185769" Sep 4 17:41:56.486050 containerd[1676]: time="2024-09-04T17:41:56.485996415Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:56.492071 containerd[1676]: time="2024-09-04T17:41:56.492017016Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:56.493584 containerd[1676]: time="2024-09-04T17:41:56.493027733Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 2.571822703s" Sep 4 17:41:56.493584 containerd[1676]: time="2024-09-04T17:41:56.493069534Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Sep 4 17:41:56.514394 containerd[1676]: time="2024-09-04T17:41:56.514351592Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Sep 4 17:41:57.145015 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2088991242.mount: Deactivated successfully. Sep 4 17:41:57.166346 containerd[1676]: time="2024-09-04T17:41:57.166296071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:57.168646 containerd[1676]: time="2024-09-04T17:41:57.168501808Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322298" Sep 4 17:41:57.175240 containerd[1676]: time="2024-09-04T17:41:57.175060118Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:57.223314 containerd[1676]: time="2024-09-04T17:41:57.223230729Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:57.225150 containerd[1676]: time="2024-09-04T17:41:57.224429850Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 710.035056ms" Sep 4 17:41:57.225150 containerd[1676]: time="2024-09-04T17:41:57.224483751Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Sep 4 17:41:57.246734 containerd[1676]: time="2024-09-04T17:41:57.246689524Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Sep 4 17:41:57.982034 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3112673449.mount: Deactivated successfully. Sep 4 17:42:00.427338 containerd[1676]: time="2024-09-04T17:42:00.427278784Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:00.429192 containerd[1676]: time="2024-09-04T17:42:00.429121715Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238579" Sep 4 17:42:00.432263 containerd[1676]: time="2024-09-04T17:42:00.432197367Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:00.438973 containerd[1676]: time="2024-09-04T17:42:00.438931280Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:00.440874 containerd[1676]: time="2024-09-04T17:42:00.440020299Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 3.193281473s" Sep 4 17:42:00.440874 containerd[1676]: time="2024-09-04T17:42:00.440062499Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" Sep 4 17:42:03.229734 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:42:03.241215 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:42:03.265217 systemd[1]: Reloading requested from client PID 2680 ('systemctl') (unit session-9.scope)... Sep 4 17:42:03.265372 systemd[1]: Reloading... Sep 4 17:42:03.387912 zram_generator::config[2714]: No configuration found. Sep 4 17:42:03.505445 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:42:03.583993 systemd[1]: Reloading finished in 318 ms. Sep 4 17:42:03.638664 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 4 17:42:03.638793 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 4 17:42:03.639214 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:42:03.640994 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:42:03.974508 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:42:03.985224 (kubelet)[2788]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:42:04.021948 kubelet[2788]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:42:04.021948 kubelet[2788]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:42:04.021948 kubelet[2788]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:42:04.689514 kubelet[2788]: I0904 17:42:04.688785 2788 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:42:04.983675 kubelet[2788]: I0904 17:42:04.983634 2788 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Sep 4 17:42:04.983675 kubelet[2788]: I0904 17:42:04.983662 2788 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:42:04.983960 kubelet[2788]: I0904 17:42:04.983939 2788 server.go:927] "Client rotation is on, will bootstrap in background" Sep 4 17:42:05.088928 kubelet[2788]: I0904 17:42:05.088162 2788 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:42:05.090186 kubelet[2788]: E0904 17:42:05.090155 2788 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.4.34:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.227951 kubelet[2788]: I0904 17:42:05.227915 2788 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:42:05.228405 kubelet[2788]: I0904 17:42:05.228212 2788 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:42:05.228550 kubelet[2788]: I0904 17:42:05.228259 2788 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4054.1.0-a-b8703b3246","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:42:05.229213 kubelet[2788]: I0904 17:42:05.229179 2788 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:42:05.229213 kubelet[2788]: I0904 17:42:05.229214 2788 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:42:05.229391 kubelet[2788]: I0904 17:42:05.229365 2788 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:42:05.287423 kubelet[2788]: I0904 17:42:05.287252 2788 kubelet.go:400] "Attempting to sync node with API server" Sep 4 17:42:05.287423 kubelet[2788]: I0904 17:42:05.287297 2788 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:42:05.287423 kubelet[2788]: I0904 17:42:05.287336 2788 kubelet.go:312] "Adding apiserver pod source" Sep 4 17:42:05.287423 kubelet[2788]: I0904 17:42:05.287359 2788 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:42:05.479813 kubelet[2788]: W0904 17:42:05.478849 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.4.34:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.479813 kubelet[2788]: E0904 17:42:05.478963 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.4.34:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.479813 kubelet[2788]: W0904 17:42:05.479050 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.479813 kubelet[2788]: E0904 17:42:05.479092 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.479813 kubelet[2788]: I0904 17:42:05.479536 2788 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.20" apiVersion="v1" Sep 4 17:42:05.482341 kubelet[2788]: I0904 17:42:05.481827 2788 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 17:42:05.482341 kubelet[2788]: W0904 17:42:05.481912 2788 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 4 17:42:05.488396 kubelet[2788]: I0904 17:42:05.488210 2788 server.go:1264] "Started kubelet" Sep 4 17:42:05.489628 kubelet[2788]: I0904 17:42:05.488615 2788 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:42:05.491385 kubelet[2788]: I0904 17:42:05.490866 2788 server.go:455] "Adding debug handlers to kubelet server" Sep 4 17:42:05.495405 kubelet[2788]: I0904 17:42:05.494361 2788 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:42:05.497904 kubelet[2788]: I0904 17:42:05.496453 2788 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 17:42:05.497904 kubelet[2788]: I0904 17:42:05.496699 2788 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:42:05.497904 kubelet[2788]: E0904 17:42:05.496851 2788 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.4.34:6443/api/v1/namespaces/default/events\": dial tcp 10.200.4.34:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4054.1.0-a-b8703b3246.17f21b6473b400ba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4054.1.0-a-b8703b3246,UID:ci-4054.1.0-a-b8703b3246,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4054.1.0-a-b8703b3246,},FirstTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,LastTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4054.1.0-a-b8703b3246,}" Sep 4 17:42:05.501444 kubelet[2788]: I0904 17:42:05.501418 2788 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:42:05.503757 kubelet[2788]: I0904 17:42:05.503725 2788 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Sep 4 17:42:05.503840 kubelet[2788]: I0904 17:42:05.503790 2788 reconciler.go:26] "Reconciler: start to sync state" Sep 4 17:42:05.515059 kubelet[2788]: E0904 17:42:05.515008 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4054.1.0-a-b8703b3246?timeout=10s\": dial tcp 10.200.4.34:6443: connect: connection refused" interval="200ms" Sep 4 17:42:05.515166 kubelet[2788]: W0904 17:42:05.515118 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.515214 kubelet[2788]: E0904 17:42:05.515181 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.525060 kubelet[2788]: I0904 17:42:05.525030 2788 factory.go:221] Registration of the containerd container factory successfully Sep 4 17:42:05.525060 kubelet[2788]: I0904 17:42:05.525056 2788 factory.go:221] Registration of the systemd container factory successfully Sep 4 17:42:05.525231 kubelet[2788]: I0904 17:42:05.525150 2788 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 17:42:05.528155 kubelet[2788]: E0904 17:42:05.527820 2788 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:42:05.595844 kubelet[2788]: I0904 17:42:05.594539 2788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:42:05.597397 kubelet[2788]: I0904 17:42:05.597044 2788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:42:05.597397 kubelet[2788]: I0904 17:42:05.597078 2788 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:42:05.597397 kubelet[2788]: I0904 17:42:05.597104 2788 kubelet.go:2337] "Starting kubelet main sync loop" Sep 4 17:42:05.597397 kubelet[2788]: E0904 17:42:05.597150 2788 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:42:05.642805 kubelet[2788]: I0904 17:42:05.642770 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.643416 kubelet[2788]: W0904 17:42:05.643167 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.643416 kubelet[2788]: E0904 17:42:05.643240 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:05.644096 kubelet[2788]: E0904 17:42:05.643736 2788 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.4.34:6443/api/v1/nodes\": dial tcp 10.200.4.34:6443: connect: connection refused" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.645094 kubelet[2788]: I0904 17:42:05.645072 2788 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:42:05.645094 kubelet[2788]: I0904 17:42:05.645085 2788 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:42:05.645205 kubelet[2788]: I0904 17:42:05.645105 2788 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:42:05.652632 kubelet[2788]: I0904 17:42:05.652532 2788 policy_none.go:49] "None policy: Start" Sep 4 17:42:05.653536 kubelet[2788]: I0904 17:42:05.653211 2788 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 4 17:42:05.653536 kubelet[2788]: I0904 17:42:05.653237 2788 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:42:05.698177 kubelet[2788]: E0904 17:42:05.698100 2788 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 17:42:05.716310 kubelet[2788]: E0904 17:42:05.716272 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4054.1.0-a-b8703b3246?timeout=10s\": dial tcp 10.200.4.34:6443: connect: connection refused" interval="400ms" Sep 4 17:42:05.789027 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 4 17:42:05.799174 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 4 17:42:05.803507 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 4 17:42:05.814767 kubelet[2788]: I0904 17:42:05.814730 2788 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:42:05.815113 kubelet[2788]: I0904 17:42:05.814989 2788 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 4 17:42:05.815170 kubelet[2788]: I0904 17:42:05.815128 2788 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:42:05.817198 kubelet[2788]: E0904 17:42:05.817147 2788 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4054.1.0-a-b8703b3246\" not found" Sep 4 17:42:05.846340 kubelet[2788]: I0904 17:42:05.846224 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.846710 kubelet[2788]: E0904 17:42:05.846647 2788 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.4.34:6443/api/v1/nodes\": dial tcp 10.200.4.34:6443: connect: connection refused" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.899020 kubelet[2788]: I0904 17:42:05.898967 2788 topology_manager.go:215] "Topology Admit Handler" podUID="53c17f29533df872577650891be90086" podNamespace="kube-system" podName="kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.900980 kubelet[2788]: I0904 17:42:05.900949 2788 topology_manager.go:215] "Topology Admit Handler" podUID="0bcf6bdfbbbf78a3745b9a8327c11de7" podNamespace="kube-system" podName="kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.903030 kubelet[2788]: I0904 17:42:05.902619 2788 topology_manager.go:215] "Topology Admit Handler" podUID="288687b7709b284245af3a1d57a60671" podNamespace="kube-system" podName="kube-scheduler-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.910130 systemd[1]: Created slice kubepods-burstable-pod53c17f29533df872577650891be90086.slice - libcontainer container kubepods-burstable-pod53c17f29533df872577650891be90086.slice. Sep 4 17:42:05.927041 systemd[1]: Created slice kubepods-burstable-pod0bcf6bdfbbbf78a3745b9a8327c11de7.slice - libcontainer container kubepods-burstable-pod0bcf6bdfbbbf78a3745b9a8327c11de7.slice. Sep 4 17:42:05.932057 systemd[1]: Created slice kubepods-burstable-pod288687b7709b284245af3a1d57a60671.slice - libcontainer container kubepods-burstable-pod288687b7709b284245af3a1d57a60671.slice. Sep 4 17:42:05.936120 kubelet[2788]: I0904 17:42:05.936018 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936366 kubelet[2788]: I0904 17:42:05.936137 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-flexvolume-dir\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936366 kubelet[2788]: I0904 17:42:05.936218 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936366 kubelet[2788]: I0904 17:42:05.936287 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-k8s-certs\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936366 kubelet[2788]: I0904 17:42:05.936358 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-kubeconfig\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936570 kubelet[2788]: I0904 17:42:05.936426 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/288687b7709b284245af3a1d57a60671-kubeconfig\") pod \"kube-scheduler-ci-4054.1.0-a-b8703b3246\" (UID: \"288687b7709b284245af3a1d57a60671\") " pod="kube-system/kube-scheduler-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936570 kubelet[2788]: I0904 17:42:05.936448 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-ca-certs\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936570 kubelet[2788]: I0904 17:42:05.936473 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-k8s-certs\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:05.936570 kubelet[2788]: I0904 17:42:05.936498 2788 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-ca-certs\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:06.117339 kubelet[2788]: E0904 17:42:06.117209 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4054.1.0-a-b8703b3246?timeout=10s\": dial tcp 10.200.4.34:6443: connect: connection refused" interval="800ms" Sep 4 17:42:06.225926 containerd[1676]: time="2024-09-04T17:42:06.225867530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4054.1.0-a-b8703b3246,Uid:53c17f29533df872577650891be90086,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:06.230600 containerd[1676]: time="2024-09-04T17:42:06.230496597Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4054.1.0-a-b8703b3246,Uid:0bcf6bdfbbbf78a3745b9a8327c11de7,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:06.235199 containerd[1676]: time="2024-09-04T17:42:06.235166966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4054.1.0-a-b8703b3246,Uid:288687b7709b284245af3a1d57a60671,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:06.249038 kubelet[2788]: I0904 17:42:06.248990 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:06.249357 kubelet[2788]: E0904 17:42:06.249327 2788 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.4.34:6443/api/v1/nodes\": dial tcp 10.200.4.34:6443: connect: connection refused" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:06.398763 kubelet[2788]: W0904 17:42:06.398627 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:06.398763 kubelet[2788]: E0904 17:42:06.398692 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:06.462957 kubelet[2788]: W0904 17:42:06.462848 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:06.462957 kubelet[2788]: E0904 17:42:06.462962 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:06.918230 kubelet[2788]: E0904 17:42:06.918169 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4054.1.0-a-b8703b3246?timeout=10s\": dial tcp 10.200.4.34:6443: connect: connection refused" interval="1.6s" Sep 4 17:42:06.958072 kubelet[2788]: E0904 17:42:06.957948 2788 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.4.34:6443/api/v1/namespaces/default/events\": dial tcp 10.200.4.34:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4054.1.0-a-b8703b3246.17f21b6473b400ba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4054.1.0-a-b8703b3246,UID:ci-4054.1.0-a-b8703b3246,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4054.1.0-a-b8703b3246,},FirstTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,LastTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4054.1.0-a-b8703b3246,}" Sep 4 17:42:07.019981 kubelet[2788]: W0904 17:42:07.019840 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.4.34:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:07.019981 kubelet[2788]: E0904 17:42:07.019932 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.4.34:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:07.051563 kubelet[2788]: I0904 17:42:07.051532 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:07.051895 kubelet[2788]: E0904 17:42:07.051853 2788 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.4.34:6443/api/v1/nodes\": dial tcp 10.200.4.34:6443: connect: connection refused" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:07.119252 kubelet[2788]: E0904 17:42:07.119203 2788 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.4.34:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:07.177088 kubelet[2788]: W0904 17:42:07.176974 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:07.177088 kubelet[2788]: E0904 17:42:07.177017 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:07.693941 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount219142690.mount: Deactivated successfully. Sep 4 17:42:07.724060 containerd[1676]: time="2024-09-04T17:42:07.724013745Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:42:07.727238 containerd[1676]: time="2024-09-04T17:42:07.727088690Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Sep 4 17:42:07.732443 containerd[1676]: time="2024-09-04T17:42:07.732402568Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:42:07.735647 containerd[1676]: time="2024-09-04T17:42:07.735613415Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:42:07.739642 containerd[1676]: time="2024-09-04T17:42:07.739593573Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:42:07.744691 containerd[1676]: time="2024-09-04T17:42:07.744648347Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:42:07.749050 containerd[1676]: time="2024-09-04T17:42:07.748785907Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:42:07.753132 containerd[1676]: time="2024-09-04T17:42:07.753085070Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:42:07.754040 containerd[1676]: time="2024-09-04T17:42:07.753797881Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.518570915s" Sep 4 17:42:07.754657 containerd[1676]: time="2024-09-04T17:42:07.754624893Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.528651462s" Sep 4 17:42:07.756338 containerd[1676]: time="2024-09-04T17:42:07.756306817Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 1.525725619s" Sep 4 17:42:08.519220 kubelet[2788]: E0904 17:42:08.519050 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.34:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4054.1.0-a-b8703b3246?timeout=10s\": dial tcp 10.200.4.34:6443: connect: connection refused" interval="3.2s" Sep 4 17:42:08.521717 containerd[1676]: time="2024-09-04T17:42:08.521022404Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:08.521717 containerd[1676]: time="2024-09-04T17:42:08.521127006Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:08.521717 containerd[1676]: time="2024-09-04T17:42:08.521163206Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.521717 containerd[1676]: time="2024-09-04T17:42:08.521307008Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.524994 containerd[1676]: time="2024-09-04T17:42:08.524637857Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:08.524994 containerd[1676]: time="2024-09-04T17:42:08.522590727Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:08.524994 containerd[1676]: time="2024-09-04T17:42:08.524866960Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:08.524994 containerd[1676]: time="2024-09-04T17:42:08.524965062Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:08.525560 containerd[1676]: time="2024-09-04T17:42:08.525141364Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.525560 containerd[1676]: time="2024-09-04T17:42:08.525253066Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.525950 containerd[1676]: time="2024-09-04T17:42:08.525709873Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.527030 containerd[1676]: time="2024-09-04T17:42:08.526976691Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:08.550310 systemd[1]: Started cri-containerd-a712571e7977840aca9961cc0734fd7c7257a00b95f82678a0cb9980e7bbc714.scope - libcontainer container a712571e7977840aca9961cc0734fd7c7257a00b95f82678a0cb9980e7bbc714. Sep 4 17:42:08.562081 systemd[1]: Started cri-containerd-3a95bda76675ce5b0cace7c68ccef0528d57cae51662219c61ccb818fe1a5782.scope - libcontainer container 3a95bda76675ce5b0cace7c68ccef0528d57cae51662219c61ccb818fe1a5782. Sep 4 17:42:08.569082 systemd[1]: Started cri-containerd-037461f7aee629651fdf582e23a4635f4c8cf76d5da19659c3e45090786d1fa8.scope - libcontainer container 037461f7aee629651fdf582e23a4635f4c8cf76d5da19659c3e45090786d1fa8. Sep 4 17:42:08.627580 kubelet[2788]: W0904 17:42:08.627366 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:08.627580 kubelet[2788]: E0904 17:42:08.627433 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.4.34:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:08.640956 kubelet[2788]: W0904 17:42:08.640799 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:08.640956 kubelet[2788]: E0904 17:42:08.640922 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.4.34:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4054.1.0-a-b8703b3246&limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:08.643216 containerd[1676]: time="2024-09-04T17:42:08.643170691Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4054.1.0-a-b8703b3246,Uid:0bcf6bdfbbbf78a3745b9a8327c11de7,Namespace:kube-system,Attempt:0,} returns sandbox id \"3a95bda76675ce5b0cace7c68ccef0528d57cae51662219c61ccb818fe1a5782\"" Sep 4 17:42:08.648816 containerd[1676]: time="2024-09-04T17:42:08.648758373Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4054.1.0-a-b8703b3246,Uid:288687b7709b284245af3a1d57a60671,Namespace:kube-system,Attempt:0,} returns sandbox id \"a712571e7977840aca9961cc0734fd7c7257a00b95f82678a0cb9980e7bbc714\"" Sep 4 17:42:08.651273 containerd[1676]: time="2024-09-04T17:42:08.651246009Z" level=info msg="CreateContainer within sandbox \"3a95bda76675ce5b0cace7c68ccef0528d57cae51662219c61ccb818fe1a5782\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 4 17:42:08.651933 containerd[1676]: time="2024-09-04T17:42:08.651879418Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4054.1.0-a-b8703b3246,Uid:53c17f29533df872577650891be90086,Namespace:kube-system,Attempt:0,} returns sandbox id \"037461f7aee629651fdf582e23a4635f4c8cf76d5da19659c3e45090786d1fa8\"" Sep 4 17:42:08.655095 containerd[1676]: time="2024-09-04T17:42:08.655067465Z" level=info msg="CreateContainer within sandbox \"a712571e7977840aca9961cc0734fd7c7257a00b95f82678a0cb9980e7bbc714\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 4 17:42:08.656541 kubelet[2788]: I0904 17:42:08.656515 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:08.657497 containerd[1676]: time="2024-09-04T17:42:08.657100895Z" level=info msg="CreateContainer within sandbox \"037461f7aee629651fdf582e23a4635f4c8cf76d5da19659c3e45090786d1fa8\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 4 17:42:08.657578 kubelet[2788]: E0904 17:42:08.657209 2788 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.4.34:6443/api/v1/nodes\": dial tcp 10.200.4.34:6443: connect: connection refused" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:08.698211 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3384342106.mount: Deactivated successfully. Sep 4 17:42:08.718328 containerd[1676]: time="2024-09-04T17:42:08.718284390Z" level=info msg="CreateContainer within sandbox \"3a95bda76675ce5b0cace7c68ccef0528d57cae51662219c61ccb818fe1a5782\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"1dfd2e336be33c49f6a26a6dda5819c6a9e5fd7c106670b9469a4a1162d331a0\"" Sep 4 17:42:08.719151 containerd[1676]: time="2024-09-04T17:42:08.718927299Z" level=info msg="StartContainer for \"1dfd2e336be33c49f6a26a6dda5819c6a9e5fd7c106670b9469a4a1162d331a0\"" Sep 4 17:42:08.733534 containerd[1676]: time="2024-09-04T17:42:08.733392111Z" level=info msg="CreateContainer within sandbox \"037461f7aee629651fdf582e23a4635f4c8cf76d5da19659c3e45090786d1fa8\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"92bbad61081865345570a22f9d0b49e0800c1ef7f9964c9d3b6f1e31a982c59b\"" Sep 4 17:42:08.734180 containerd[1676]: time="2024-09-04T17:42:08.734148622Z" level=info msg="StartContainer for \"92bbad61081865345570a22f9d0b49e0800c1ef7f9964c9d3b6f1e31a982c59b\"" Sep 4 17:42:08.736919 containerd[1676]: time="2024-09-04T17:42:08.735863947Z" level=info msg="CreateContainer within sandbox \"a712571e7977840aca9961cc0734fd7c7257a00b95f82678a0cb9980e7bbc714\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f1bea913479eb3896ae584c7773870a9e68e019e31b68449798e7bad2ebbc6c7\"" Sep 4 17:42:08.737851 containerd[1676]: time="2024-09-04T17:42:08.737827076Z" level=info msg="StartContainer for \"f1bea913479eb3896ae584c7773870a9e68e019e31b68449798e7bad2ebbc6c7\"" Sep 4 17:42:08.746096 systemd[1]: Started cri-containerd-1dfd2e336be33c49f6a26a6dda5819c6a9e5fd7c106670b9469a4a1162d331a0.scope - libcontainer container 1dfd2e336be33c49f6a26a6dda5819c6a9e5fd7c106670b9469a4a1162d331a0. Sep 4 17:42:08.785109 systemd[1]: Started cri-containerd-92bbad61081865345570a22f9d0b49e0800c1ef7f9964c9d3b6f1e31a982c59b.scope - libcontainer container 92bbad61081865345570a22f9d0b49e0800c1ef7f9964c9d3b6f1e31a982c59b. Sep 4 17:42:08.801018 systemd[1]: Started cri-containerd-f1bea913479eb3896ae584c7773870a9e68e019e31b68449798e7bad2ebbc6c7.scope - libcontainer container f1bea913479eb3896ae584c7773870a9e68e019e31b68449798e7bad2ebbc6c7. Sep 4 17:42:08.827918 containerd[1676]: time="2024-09-04T17:42:08.827866493Z" level=info msg="StartContainer for \"1dfd2e336be33c49f6a26a6dda5819c6a9e5fd7c106670b9469a4a1162d331a0\" returns successfully" Sep 4 17:42:08.887908 containerd[1676]: time="2024-09-04T17:42:08.886995258Z" level=info msg="StartContainer for \"92bbad61081865345570a22f9d0b49e0800c1ef7f9964c9d3b6f1e31a982c59b\" returns successfully" Sep 4 17:42:08.904564 containerd[1676]: time="2024-09-04T17:42:08.904504314Z" level=info msg="StartContainer for \"f1bea913479eb3896ae584c7773870a9e68e019e31b68449798e7bad2ebbc6c7\" returns successfully" Sep 4 17:42:08.973188 kubelet[2788]: W0904 17:42:08.973107 2788 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:08.973366 kubelet[2788]: E0904 17:42:08.973203 2788 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.4.34:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.34:6443: connect: connection refused Sep 4 17:42:11.718821 kubelet[2788]: E0904 17:42:11.718777 2788 csi_plugin.go:308] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "ci-4054.1.0-a-b8703b3246" not found Sep 4 17:42:11.722331 kubelet[2788]: E0904 17:42:11.722280 2788 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4054.1.0-a-b8703b3246\" not found" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:11.859582 kubelet[2788]: I0904 17:42:11.859536 2788 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:13.085874 kubelet[2788]: I0904 17:42:13.084281 2788 kubelet_node_status.go:76] "Successfully registered node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:13.724976 systemd[1]: Reloading requested from client PID 3064 ('systemctl') (unit session-9.scope)... Sep 4 17:42:13.724991 systemd[1]: Reloading... Sep 4 17:42:13.825919 zram_generator::config[3101]: No configuration found. Sep 4 17:42:13.948801 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:42:14.490584 kubelet[2788]: I0904 17:42:14.082398 2788 apiserver.go:52] "Watching apiserver" Sep 4 17:42:14.490584 kubelet[2788]: E0904 17:42:14.084232 2788 event.go:319] "Unable to write event (broadcaster is shut down)" event="&Event{ObjectMeta:{ci-4054.1.0-a-b8703b3246.17f21b6473b400ba default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4054.1.0-a-b8703b3246,UID:ci-4054.1.0-a-b8703b3246,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4054.1.0-a-b8703b3246,},FirstTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,LastTimestamp:2024-09-04 17:42:05.488177338 +0000 UTC m=+1.499673139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4054.1.0-a-b8703b3246,}" Sep 4 17:42:14.041860 systemd[1]: Reloading finished in 316 ms. Sep 4 17:42:14.084394 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:42:14.098293 systemd[1]: kubelet.service: Deactivated successfully. Sep 4 17:42:14.098510 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:42:14.105161 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:42:14.611042 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:42:14.623227 (kubelet)[3168]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:42:14.661665 kubelet[3168]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:42:14.661665 kubelet[3168]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:42:14.661665 kubelet[3168]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:42:14.662172 kubelet[3168]: I0904 17:42:14.661722 3168 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:42:14.667917 kubelet[3168]: I0904 17:42:14.667632 3168 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Sep 4 17:42:14.667917 kubelet[3168]: I0904 17:42:14.667654 3168 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:42:14.667917 kubelet[3168]: I0904 17:42:14.667812 3168 server.go:927] "Client rotation is on, will bootstrap in background" Sep 4 17:42:14.669021 kubelet[3168]: I0904 17:42:14.668994 3168 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 4 17:42:14.670211 kubelet[3168]: I0904 17:42:14.670067 3168 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:42:14.676340 kubelet[3168]: I0904 17:42:14.676305 3168 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:42:14.676597 kubelet[3168]: I0904 17:42:14.676546 3168 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:42:14.676774 kubelet[3168]: I0904 17:42:14.676591 3168 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4054.1.0-a-b8703b3246","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:42:14.676923 kubelet[3168]: I0904 17:42:14.676787 3168 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:42:14.676923 kubelet[3168]: I0904 17:42:14.676802 3168 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:42:14.676923 kubelet[3168]: I0904 17:42:14.676856 3168 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:42:14.677058 kubelet[3168]: I0904 17:42:14.676991 3168 kubelet.go:400] "Attempting to sync node with API server" Sep 4 17:42:14.677058 kubelet[3168]: I0904 17:42:14.677006 3168 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:42:14.677058 kubelet[3168]: I0904 17:42:14.677031 3168 kubelet.go:312] "Adding apiserver pod source" Sep 4 17:42:14.677058 kubelet[3168]: I0904 17:42:14.677050 3168 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:42:14.681468 kubelet[3168]: I0904 17:42:14.680149 3168 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.20" apiVersion="v1" Sep 4 17:42:14.681468 kubelet[3168]: I0904 17:42:14.680379 3168 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 17:42:14.681468 kubelet[3168]: I0904 17:42:14.680822 3168 server.go:1264] "Started kubelet" Sep 4 17:42:14.686122 kubelet[3168]: I0904 17:42:14.686106 3168 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:42:14.693273 kubelet[3168]: I0904 17:42:14.693244 3168 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:42:14.696616 kubelet[3168]: I0904 17:42:14.696588 3168 server.go:455] "Adding debug handlers to kubelet server" Sep 4 17:42:14.699111 kubelet[3168]: I0904 17:42:14.699048 3168 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 17:42:14.699528 kubelet[3168]: I0904 17:42:14.699505 3168 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:42:14.701833 kubelet[3168]: I0904 17:42:14.701814 3168 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:42:14.706367 kubelet[3168]: I0904 17:42:14.706344 3168 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Sep 4 17:42:14.706613 kubelet[3168]: I0904 17:42:14.706505 3168 reconciler.go:26] "Reconciler: start to sync state" Sep 4 17:42:14.712237 kubelet[3168]: I0904 17:42:14.712219 3168 factory.go:221] Registration of the systemd container factory successfully Sep 4 17:42:14.712452 kubelet[3168]: I0904 17:42:14.712433 3168 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 17:42:14.713358 kubelet[3168]: I0904 17:42:14.713328 3168 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:42:14.715789 kubelet[3168]: I0904 17:42:14.715768 3168 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:42:14.716385 kubelet[3168]: I0904 17:42:14.715926 3168 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:42:14.716385 kubelet[3168]: I0904 17:42:14.715953 3168 kubelet.go:2337] "Starting kubelet main sync loop" Sep 4 17:42:14.716385 kubelet[3168]: E0904 17:42:14.715999 3168 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:42:14.721643 kubelet[3168]: I0904 17:42:14.721577 3168 factory.go:221] Registration of the containerd container factory successfully Sep 4 17:42:14.723986 kubelet[3168]: E0904 17:42:14.723585 3168 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:42:14.768673 kubelet[3168]: I0904 17:42:14.768583 3168 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:42:14.768922 kubelet[3168]: I0904 17:42:14.768904 3168 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:42:14.769040 kubelet[3168]: I0904 17:42:14.769030 3168 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:42:14.769386 kubelet[3168]: I0904 17:42:14.769254 3168 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 4 17:42:14.769386 kubelet[3168]: I0904 17:42:14.769268 3168 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 4 17:42:14.769386 kubelet[3168]: I0904 17:42:14.769303 3168 policy_none.go:49] "None policy: Start" Sep 4 17:42:14.770625 kubelet[3168]: I0904 17:42:14.770605 3168 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 4 17:42:14.770702 kubelet[3168]: I0904 17:42:14.770632 3168 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:42:14.771120 kubelet[3168]: I0904 17:42:14.770860 3168 state_mem.go:75] "Updated machine memory state" Sep 4 17:42:14.774996 kubelet[3168]: I0904 17:42:14.774981 3168 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:42:14.775399 kubelet[3168]: I0904 17:42:14.775242 3168 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 4 17:42:14.775399 kubelet[3168]: I0904 17:42:14.775346 3168 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:42:14.810439 kubelet[3168]: I0904 17:42:14.809123 3168 kubelet_node_status.go:73] "Attempting to register node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:14.816447 kubelet[3168]: I0904 17:42:14.816410 3168 topology_manager.go:215] "Topology Admit Handler" podUID="53c17f29533df872577650891be90086" podNamespace="kube-system" podName="kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:14.816670 kubelet[3168]: I0904 17:42:14.816512 3168 topology_manager.go:215] "Topology Admit Handler" podUID="0bcf6bdfbbbf78a3745b9a8327c11de7" podNamespace="kube-system" podName="kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:14.816670 kubelet[3168]: I0904 17:42:14.816582 3168 topology_manager.go:215] "Topology Admit Handler" podUID="288687b7709b284245af3a1d57a60671" podNamespace="kube-system" podName="kube-scheduler-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186329 kubelet[3168]: I0904 17:42:15.184811 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-ca-certs\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186329 kubelet[3168]: I0904 17:42:15.184922 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-k8s-certs\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186329 kubelet[3168]: I0904 17:42:15.184953 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/288687b7709b284245af3a1d57a60671-kubeconfig\") pod \"kube-scheduler-ci-4054.1.0-a-b8703b3246\" (UID: \"288687b7709b284245af3a1d57a60671\") " pod="kube-system/kube-scheduler-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186329 kubelet[3168]: I0904 17:42:15.184976 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/53c17f29533df872577650891be90086-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" (UID: \"53c17f29533df872577650891be90086\") " pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186329 kubelet[3168]: I0904 17:42:15.185002 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-ca-certs\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186963 kubelet[3168]: I0904 17:42:15.185024 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-flexvolume-dir\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186963 kubelet[3168]: I0904 17:42:15.185045 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-k8s-certs\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186963 kubelet[3168]: I0904 17:42:15.185067 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-kubeconfig\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.186963 kubelet[3168]: I0904 17:42:15.185091 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0bcf6bdfbbbf78a3745b9a8327c11de7-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4054.1.0-a-b8703b3246\" (UID: \"0bcf6bdfbbbf78a3745b9a8327c11de7\") " pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.192349 kubelet[3168]: I0904 17:42:15.191836 3168 kubelet_node_status.go:112] "Node was previously registered" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.192349 kubelet[3168]: I0904 17:42:15.191938 3168 kubelet_node_status.go:76] "Successfully registered node" node="ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.196678 kubelet[3168]: W0904 17:42:15.196628 3168 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 4 17:42:15.197828 kubelet[3168]: W0904 17:42:15.197592 3168 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 4 17:42:15.197828 kubelet[3168]: W0904 17:42:15.197730 3168 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 4 17:42:15.677762 kubelet[3168]: I0904 17:42:15.677529 3168 apiserver.go:52] "Watching apiserver" Sep 4 17:42:15.707566 kubelet[3168]: I0904 17:42:15.707499 3168 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Sep 4 17:42:15.779673 kubelet[3168]: W0904 17:42:15.778754 3168 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Sep 4 17:42:15.779673 kubelet[3168]: E0904 17:42:15.779076 3168 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4054.1.0-a-b8703b3246\" already exists" pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" Sep 4 17:42:15.790775 kubelet[3168]: I0904 17:42:15.790594 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4054.1.0-a-b8703b3246" podStartSLOduration=0.790469485 podStartE2EDuration="790.469485ms" podCreationTimestamp="2024-09-04 17:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:42:15.7789216 +0000 UTC m=+1.152273042" watchObservedRunningTime="2024-09-04 17:42:15.790469485 +0000 UTC m=+1.163820827" Sep 4 17:42:15.791005 kubelet[3168]: I0904 17:42:15.790860 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4054.1.0-a-b8703b3246" podStartSLOduration=0.790848191 podStartE2EDuration="790.848191ms" podCreationTimestamp="2024-09-04 17:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:42:15.789589971 +0000 UTC m=+1.162941313" watchObservedRunningTime="2024-09-04 17:42:15.790848191 +0000 UTC m=+1.164199533" Sep 4 17:42:15.798030 kubelet[3168]: I0904 17:42:15.797759 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4054.1.0-a-b8703b3246" podStartSLOduration=0.797741201 podStartE2EDuration="797.741201ms" podCreationTimestamp="2024-09-04 17:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:42:15.797564298 +0000 UTC m=+1.170915640" watchObservedRunningTime="2024-09-04 17:42:15.797741201 +0000 UTC m=+1.171092743" Sep 4 17:42:19.834427 sudo[2209]: pam_unix(sudo:session): session closed for user root Sep 4 17:42:19.927646 sshd[2206]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:19.932524 systemd[1]: sshd@6-10.200.4.34:22-10.200.16.10:51878.service: Deactivated successfully. Sep 4 17:42:19.934488 systemd[1]: session-9.scope: Deactivated successfully. Sep 4 17:42:19.934700 systemd[1]: session-9.scope: Consumed 4.778s CPU time, 141.2M memory peak, 0B memory swap peak. Sep 4 17:42:19.935420 systemd-logind[1648]: Session 9 logged out. Waiting for processes to exit. Sep 4 17:42:19.936412 systemd-logind[1648]: Removed session 9. Sep 4 17:42:27.932495 kubelet[3168]: I0904 17:42:27.932385 3168 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 4 17:42:27.933227 kubelet[3168]: I0904 17:42:27.933082 3168 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 4 17:42:27.933280 containerd[1676]: time="2024-09-04T17:42:27.932841654Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 4 17:42:28.787157 kubelet[3168]: I0904 17:42:28.785672 3168 topology_manager.go:215] "Topology Admit Handler" podUID="54b91def-8d3a-4188-bda1-ada4848211b4" podNamespace="kube-system" podName="kube-proxy-mtqtn" Sep 4 17:42:28.797718 systemd[1]: Created slice kubepods-besteffort-pod54b91def_8d3a_4188_bda1_ada4848211b4.slice - libcontainer container kubepods-besteffort-pod54b91def_8d3a_4188_bda1_ada4848211b4.slice. Sep 4 17:42:28.881296 kubelet[3168]: I0904 17:42:28.881256 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/54b91def-8d3a-4188-bda1-ada4848211b4-kube-proxy\") pod \"kube-proxy-mtqtn\" (UID: \"54b91def-8d3a-4188-bda1-ada4848211b4\") " pod="kube-system/kube-proxy-mtqtn" Sep 4 17:42:28.881296 kubelet[3168]: I0904 17:42:28.881321 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/54b91def-8d3a-4188-bda1-ada4848211b4-lib-modules\") pod \"kube-proxy-mtqtn\" (UID: \"54b91def-8d3a-4188-bda1-ada4848211b4\") " pod="kube-system/kube-proxy-mtqtn" Sep 4 17:42:28.881296 kubelet[3168]: I0904 17:42:28.881351 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7jkt\" (UniqueName: \"kubernetes.io/projected/54b91def-8d3a-4188-bda1-ada4848211b4-kube-api-access-q7jkt\") pod \"kube-proxy-mtqtn\" (UID: \"54b91def-8d3a-4188-bda1-ada4848211b4\") " pod="kube-system/kube-proxy-mtqtn" Sep 4 17:42:28.881582 kubelet[3168]: I0904 17:42:28.881377 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/54b91def-8d3a-4188-bda1-ada4848211b4-xtables-lock\") pod \"kube-proxy-mtqtn\" (UID: \"54b91def-8d3a-4188-bda1-ada4848211b4\") " pod="kube-system/kube-proxy-mtqtn" Sep 4 17:42:28.957104 kubelet[3168]: I0904 17:42:28.956186 3168 topology_manager.go:215] "Topology Admit Handler" podUID="c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4" podNamespace="tigera-operator" podName="tigera-operator-77f994b5bb-r4c22" Sep 4 17:42:28.970123 systemd[1]: Created slice kubepods-besteffort-podc2066b46_5f9a_44fd_a2ee_18f5a5cd65a4.slice - libcontainer container kubepods-besteffort-podc2066b46_5f9a_44fd_a2ee_18f5a5cd65a4.slice. Sep 4 17:42:29.082422 kubelet[3168]: I0904 17:42:29.082280 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsvx\" (UniqueName: \"kubernetes.io/projected/c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4-kube-api-access-bcsvx\") pod \"tigera-operator-77f994b5bb-r4c22\" (UID: \"c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4\") " pod="tigera-operator/tigera-operator-77f994b5bb-r4c22" Sep 4 17:42:29.082422 kubelet[3168]: I0904 17:42:29.082322 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4-var-lib-calico\") pod \"tigera-operator-77f994b5bb-r4c22\" (UID: \"c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4\") " pod="tigera-operator/tigera-operator-77f994b5bb-r4c22" Sep 4 17:42:29.106258 containerd[1676]: time="2024-09-04T17:42:29.106208578Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-mtqtn,Uid:54b91def-8d3a-4188-bda1-ada4848211b4,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:29.193649 containerd[1676]: time="2024-09-04T17:42:29.193303853Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:29.193649 containerd[1676]: time="2024-09-04T17:42:29.193424655Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:29.193649 containerd[1676]: time="2024-09-04T17:42:29.193442855Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:29.198193 containerd[1676]: time="2024-09-04T17:42:29.193769960Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:29.223058 systemd[1]: Started cri-containerd-2d31ca3d803fe3461e0d9ad2e7fc6691a3bd0e287cb3d888e7fd317be83d0539.scope - libcontainer container 2d31ca3d803fe3461e0d9ad2e7fc6691a3bd0e287cb3d888e7fd317be83d0539. Sep 4 17:42:29.245517 containerd[1676]: time="2024-09-04T17:42:29.245477776Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-mtqtn,Uid:54b91def-8d3a-4188-bda1-ada4848211b4,Namespace:kube-system,Attempt:0,} returns sandbox id \"2d31ca3d803fe3461e0d9ad2e7fc6691a3bd0e287cb3d888e7fd317be83d0539\"" Sep 4 17:42:29.248589 containerd[1676]: time="2024-09-04T17:42:29.248490324Z" level=info msg="CreateContainer within sandbox \"2d31ca3d803fe3461e0d9ad2e7fc6691a3bd0e287cb3d888e7fd317be83d0539\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 4 17:42:29.273985 containerd[1676]: time="2024-09-04T17:42:29.273677722Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-77f994b5bb-r4c22,Uid:c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4,Namespace:tigera-operator,Attempt:0,}" Sep 4 17:42:29.281315 containerd[1676]: time="2024-09-04T17:42:29.281239141Z" level=info msg="CreateContainer within sandbox \"2d31ca3d803fe3461e0d9ad2e7fc6691a3bd0e287cb3d888e7fd317be83d0539\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9a5689e099bb18d30e4a4aa6a62e73caf6aeefb7aa4b2d0485dbca2f413796b3\"" Sep 4 17:42:29.281964 containerd[1676]: time="2024-09-04T17:42:29.281932752Z" level=info msg="StartContainer for \"9a5689e099bb18d30e4a4aa6a62e73caf6aeefb7aa4b2d0485dbca2f413796b3\"" Sep 4 17:42:29.311361 systemd[1]: Started cri-containerd-9a5689e099bb18d30e4a4aa6a62e73caf6aeefb7aa4b2d0485dbca2f413796b3.scope - libcontainer container 9a5689e099bb18d30e4a4aa6a62e73caf6aeefb7aa4b2d0485dbca2f413796b3. Sep 4 17:42:29.331193 containerd[1676]: time="2024-09-04T17:42:29.331086728Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:29.331530 containerd[1676]: time="2024-09-04T17:42:29.331250631Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:29.331530 containerd[1676]: time="2024-09-04T17:42:29.331368632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:29.334219 containerd[1676]: time="2024-09-04T17:42:29.332527951Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:29.356104 systemd[1]: Started cri-containerd-013e91c2fcbaedd886e930e58edfa87b285caeebe84b12e70161498579c6e93e.scope - libcontainer container 013e91c2fcbaedd886e930e58edfa87b285caeebe84b12e70161498579c6e93e. Sep 4 17:42:29.366687 containerd[1676]: time="2024-09-04T17:42:29.366404886Z" level=info msg="StartContainer for \"9a5689e099bb18d30e4a4aa6a62e73caf6aeefb7aa4b2d0485dbca2f413796b3\" returns successfully" Sep 4 17:42:29.419668 containerd[1676]: time="2024-09-04T17:42:29.419616826Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-77f994b5bb-r4c22,Uid:c2066b46-5f9a-44fd-a2ee-18f5a5cd65a4,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"013e91c2fcbaedd886e930e58edfa87b285caeebe84b12e70161498579c6e93e\"" Sep 4 17:42:29.423853 containerd[1676]: time="2024-09-04T17:42:29.423816992Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.3\"" Sep 4 17:42:29.789215 kubelet[3168]: I0904 17:42:29.788837 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-mtqtn" podStartSLOduration=1.7888182540000002 podStartE2EDuration="1.788818254s" podCreationTimestamp="2024-09-04 17:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:42:29.788710252 +0000 UTC m=+15.162061694" watchObservedRunningTime="2024-09-04 17:42:29.788818254 +0000 UTC m=+15.162169596" Sep 4 17:42:31.172437 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount814059932.mount: Deactivated successfully. Sep 4 17:42:31.737611 containerd[1676]: time="2024-09-04T17:42:31.737500618Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:31.740076 containerd[1676]: time="2024-09-04T17:42:31.739933857Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.34.3: active requests=0, bytes read=22136541" Sep 4 17:42:31.745119 containerd[1676]: time="2024-09-04T17:42:31.745062338Z" level=info msg="ImageCreate event name:\"sha256:d4e6e064c25d51e66b2470e80d7b57004f79e2a76b37e83986577f8666da9736\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:31.749377 containerd[1676]: time="2024-09-04T17:42:31.749266204Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:2cc4de6ad019ccc3abbd2615c159d0dcfb2ecdab90dc5805f08837d7c014d458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:31.750722 containerd[1676]: time="2024-09-04T17:42:31.750423422Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.34.3\" with image id \"sha256:d4e6e064c25d51e66b2470e80d7b57004f79e2a76b37e83986577f8666da9736\", repo tag \"quay.io/tigera/operator:v1.34.3\", repo digest \"quay.io/tigera/operator@sha256:2cc4de6ad019ccc3abbd2615c159d0dcfb2ecdab90dc5805f08837d7c014d458\", size \"22130728\" in 2.326464028s" Sep 4 17:42:31.750722 containerd[1676]: time="2024-09-04T17:42:31.750462823Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.3\" returns image reference \"sha256:d4e6e064c25d51e66b2470e80d7b57004f79e2a76b37e83986577f8666da9736\"" Sep 4 17:42:31.753246 containerd[1676]: time="2024-09-04T17:42:31.753114665Z" level=info msg="CreateContainer within sandbox \"013e91c2fcbaedd886e930e58edfa87b285caeebe84b12e70161498579c6e93e\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 4 17:42:31.791773 containerd[1676]: time="2024-09-04T17:42:31.791731074Z" level=info msg="CreateContainer within sandbox \"013e91c2fcbaedd886e930e58edfa87b285caeebe84b12e70161498579c6e93e\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"defe2304d041cd675501264071b53ece0e2ecee2343785ac628ab7a393458e9a\"" Sep 4 17:42:31.792918 containerd[1676]: time="2024-09-04T17:42:31.792216482Z" level=info msg="StartContainer for \"defe2304d041cd675501264071b53ece0e2ecee2343785ac628ab7a393458e9a\"" Sep 4 17:42:31.820559 systemd[1]: Started cri-containerd-defe2304d041cd675501264071b53ece0e2ecee2343785ac628ab7a393458e9a.scope - libcontainer container defe2304d041cd675501264071b53ece0e2ecee2343785ac628ab7a393458e9a. Sep 4 17:42:31.844580 containerd[1676]: time="2024-09-04T17:42:31.844531008Z" level=info msg="StartContainer for \"defe2304d041cd675501264071b53ece0e2ecee2343785ac628ab7a393458e9a\" returns successfully" Sep 4 17:42:32.796169 kubelet[3168]: I0904 17:42:32.795831 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-77f994b5bb-r4c22" podStartSLOduration=2.465900343 podStartE2EDuration="4.795810626s" podCreationTimestamp="2024-09-04 17:42:28 +0000 UTC" firstStartedPulling="2024-09-04 17:42:29.421683858 +0000 UTC m=+14.795035200" lastFinishedPulling="2024-09-04 17:42:31.751594141 +0000 UTC m=+17.124945483" observedRunningTime="2024-09-04 17:42:32.795748625 +0000 UTC m=+18.169099967" watchObservedRunningTime="2024-09-04 17:42:32.795810626 +0000 UTC m=+18.169162068" Sep 4 17:42:34.706907 kubelet[3168]: I0904 17:42:34.706837 3168 topology_manager.go:215] "Topology Admit Handler" podUID="cb36bc54-36ce-4298-94ef-03febcd650bf" podNamespace="calico-system" podName="calico-typha-675d96d689-vxm87" Sep 4 17:42:34.717906 systemd[1]: Created slice kubepods-besteffort-podcb36bc54_36ce_4298_94ef_03febcd650bf.slice - libcontainer container kubepods-besteffort-podcb36bc54_36ce_4298_94ef_03febcd650bf.slice. Sep 4 17:42:34.720565 kubelet[3168]: W0904 17:42:34.719875 3168 reflector.go:547] object-"calico-system"/"typha-certs": failed to list *v1.Secret: secrets "typha-certs" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.720565 kubelet[3168]: E0904 17:42:34.719942 3168 reflector.go:150] object-"calico-system"/"typha-certs": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "typha-certs" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.720565 kubelet[3168]: W0904 17:42:34.720001 3168 reflector.go:547] object-"calico-system"/"tigera-ca-bundle": failed to list *v1.ConfigMap: configmaps "tigera-ca-bundle" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.720565 kubelet[3168]: E0904 17:42:34.720013 3168 reflector.go:150] object-"calico-system"/"tigera-ca-bundle": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "tigera-ca-bundle" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.720565 kubelet[3168]: W0904 17:42:34.720050 3168 reflector.go:547] object-"calico-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.721336 kubelet[3168]: E0904 17:42:34.720061 3168 reflector.go:150] object-"calico-system"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:42:34.822597 kubelet[3168]: I0904 17:42:34.822028 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb36bc54-36ce-4298-94ef-03febcd650bf-tigera-ca-bundle\") pod \"calico-typha-675d96d689-vxm87\" (UID: \"cb36bc54-36ce-4298-94ef-03febcd650bf\") " pod="calico-system/calico-typha-675d96d689-vxm87" Sep 4 17:42:34.822597 kubelet[3168]: I0904 17:42:34.822209 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/cb36bc54-36ce-4298-94ef-03febcd650bf-typha-certs\") pod \"calico-typha-675d96d689-vxm87\" (UID: \"cb36bc54-36ce-4298-94ef-03febcd650bf\") " pod="calico-system/calico-typha-675d96d689-vxm87" Sep 4 17:42:34.822597 kubelet[3168]: I0904 17:42:34.822290 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fc4k\" (UniqueName: \"kubernetes.io/projected/cb36bc54-36ce-4298-94ef-03febcd650bf-kube-api-access-8fc4k\") pod \"calico-typha-675d96d689-vxm87\" (UID: \"cb36bc54-36ce-4298-94ef-03febcd650bf\") " pod="calico-system/calico-typha-675d96d689-vxm87" Sep 4 17:42:34.894457 kubelet[3168]: I0904 17:42:34.894406 3168 topology_manager.go:215] "Topology Admit Handler" podUID="2f2ca5b5-17c8-43a2-9515-098a3c51c0f8" podNamespace="calico-system" podName="calico-node-rcds4" Sep 4 17:42:34.904067 systemd[1]: Created slice kubepods-besteffort-pod2f2ca5b5_17c8_43a2_9515_098a3c51c0f8.slice - libcontainer container kubepods-besteffort-pod2f2ca5b5_17c8_43a2_9515_098a3c51c0f8.slice. Sep 4 17:42:35.011623 kubelet[3168]: I0904 17:42:35.011522 3168 topology_manager.go:215] "Topology Admit Handler" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" podNamespace="calico-system" podName="csi-node-driver-msd5w" Sep 4 17:42:35.012735 kubelet[3168]: E0904 17:42:35.012467 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:35.024384 kubelet[3168]: I0904 17:42:35.023506 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-cni-bin-dir\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024384 kubelet[3168]: I0904 17:42:35.023551 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-cni-log-dir\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024384 kubelet[3168]: I0904 17:42:35.023577 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-lib-modules\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024384 kubelet[3168]: I0904 17:42:35.023614 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-xtables-lock\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024384 kubelet[3168]: I0904 17:42:35.023641 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-policysync\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024660 kubelet[3168]: I0904 17:42:35.023667 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-var-run-calico\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024660 kubelet[3168]: I0904 17:42:35.023694 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-flexvol-driver-host\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024660 kubelet[3168]: I0904 17:42:35.023719 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lfrf\" (UniqueName: \"kubernetes.io/projected/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-kube-api-access-7lfrf\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024660 kubelet[3168]: I0904 17:42:35.023745 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-tigera-ca-bundle\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024660 kubelet[3168]: I0904 17:42:35.023770 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-node-certs\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024857 kubelet[3168]: I0904 17:42:35.023798 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-cni-net-dir\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.024857 kubelet[3168]: I0904 17:42:35.023831 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/2f2ca5b5-17c8-43a2-9515-098a3c51c0f8-var-lib-calico\") pod \"calico-node-rcds4\" (UID: \"2f2ca5b5-17c8-43a2-9515-098a3c51c0f8\") " pod="calico-system/calico-node-rcds4" Sep 4 17:42:35.125082 kubelet[3168]: I0904 17:42:35.124668 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d-registration-dir\") pod \"csi-node-driver-msd5w\" (UID: \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\") " pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:35.125082 kubelet[3168]: I0904 17:42:35.124744 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d-socket-dir\") pod \"csi-node-driver-msd5w\" (UID: \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\") " pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:35.125082 kubelet[3168]: I0904 17:42:35.124918 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m98v\" (UniqueName: \"kubernetes.io/projected/a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d-kube-api-access-5m98v\") pod \"csi-node-driver-msd5w\" (UID: \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\") " pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:35.125082 kubelet[3168]: I0904 17:42:35.124991 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d-varrun\") pod \"csi-node-driver-msd5w\" (UID: \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\") " pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:35.125082 kubelet[3168]: I0904 17:42:35.125012 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d-kubelet-dir\") pod \"csi-node-driver-msd5w\" (UID: \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\") " pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:35.129004 kubelet[3168]: E0904 17:42:35.128970 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.129004 kubelet[3168]: W0904 17:42:35.129002 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.129151 kubelet[3168]: E0904 17:42:35.129027 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.129603 kubelet[3168]: E0904 17:42:35.129362 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.129603 kubelet[3168]: W0904 17:42:35.129378 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.129603 kubelet[3168]: E0904 17:42:35.129396 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.132449 kubelet[3168]: E0904 17:42:35.132434 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.132758 kubelet[3168]: W0904 17:42:35.132548 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.132758 kubelet[3168]: E0904 17:42:35.132570 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.135177 kubelet[3168]: E0904 17:42:35.133382 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.135177 kubelet[3168]: W0904 17:42:35.133399 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.135177 kubelet[3168]: E0904 17:42:35.133415 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.135488 kubelet[3168]: E0904 17:42:35.135393 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.135488 kubelet[3168]: W0904 17:42:35.135407 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.135675 kubelet[3168]: E0904 17:42:35.135603 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.135815 kubelet[3168]: E0904 17:42:35.135804 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.135948 kubelet[3168]: W0904 17:42:35.135879 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.136117 kubelet[3168]: E0904 17:42:35.136024 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.136337 kubelet[3168]: E0904 17:42:35.136288 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.136337 kubelet[3168]: W0904 17:42:35.136318 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.136615 kubelet[3168]: E0904 17:42:35.136532 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.137078 kubelet[3168]: E0904 17:42:35.137063 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.137983 kubelet[3168]: W0904 17:42:35.137165 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.138157 kubelet[3168]: E0904 17:42:35.138077 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.139410 kubelet[3168]: E0904 17:42:35.139332 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.139410 kubelet[3168]: W0904 17:42:35.139347 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.139776 kubelet[3168]: E0904 17:42:35.139556 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.140169 kubelet[3168]: E0904 17:42:35.140005 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.140169 kubelet[3168]: W0904 17:42:35.140019 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.140598 kubelet[3168]: E0904 17:42:35.140366 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.140598 kubelet[3168]: E0904 17:42:35.140572 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.140598 kubelet[3168]: W0904 17:42:35.140583 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.141090 kubelet[3168]: E0904 17:42:35.140848 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.141418 kubelet[3168]: E0904 17:42:35.141370 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.141418 kubelet[3168]: W0904 17:42:35.141383 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.141763 kubelet[3168]: E0904 17:42:35.141746 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.142200 kubelet[3168]: E0904 17:42:35.142177 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.142346 kubelet[3168]: W0904 17:42:35.142197 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.142346 kubelet[3168]: E0904 17:42:35.142319 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.142545 kubelet[3168]: E0904 17:42:35.142532 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.142765 kubelet[3168]: W0904 17:42:35.142642 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.142765 kubelet[3168]: E0904 17:42:35.142687 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.143321 kubelet[3168]: E0904 17:42:35.143257 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.143321 kubelet[3168]: W0904 17:42:35.143274 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.143321 kubelet[3168]: E0904 17:42:35.143289 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.226277 kubelet[3168]: E0904 17:42:35.226000 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.226277 kubelet[3168]: W0904 17:42:35.226028 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.226277 kubelet[3168]: E0904 17:42:35.226057 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.226906 kubelet[3168]: E0904 17:42:35.226785 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.226906 kubelet[3168]: W0904 17:42:35.226805 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.226906 kubelet[3168]: E0904 17:42:35.226840 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.227591 kubelet[3168]: E0904 17:42:35.227342 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.227591 kubelet[3168]: W0904 17:42:35.227357 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.227591 kubelet[3168]: E0904 17:42:35.227387 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.227852 kubelet[3168]: E0904 17:42:35.227796 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.227852 kubelet[3168]: W0904 17:42:35.227811 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.227852 kubelet[3168]: E0904 17:42:35.227841 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.228350 kubelet[3168]: E0904 17:42:35.228105 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.228350 kubelet[3168]: W0904 17:42:35.228119 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.228350 kubelet[3168]: E0904 17:42:35.228208 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.228503 kubelet[3168]: E0904 17:42:35.228398 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.228503 kubelet[3168]: W0904 17:42:35.228409 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.228503 kubelet[3168]: E0904 17:42:35.228499 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.228643 kubelet[3168]: E0904 17:42:35.228627 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.228643 kubelet[3168]: W0904 17:42:35.228635 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.228723 kubelet[3168]: E0904 17:42:35.228663 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.229210 kubelet[3168]: E0904 17:42:35.228895 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.229210 kubelet[3168]: W0904 17:42:35.228911 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.229210 kubelet[3168]: E0904 17:42:35.228929 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.229210 kubelet[3168]: E0904 17:42:35.229168 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.229210 kubelet[3168]: W0904 17:42:35.229180 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.229210 kubelet[3168]: E0904 17:42:35.229206 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.229588 kubelet[3168]: E0904 17:42:35.229480 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.229588 kubelet[3168]: W0904 17:42:35.229492 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.229588 kubelet[3168]: E0904 17:42:35.229575 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230021 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.230724 kubelet[3168]: W0904 17:42:35.230037 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230125 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230284 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.230724 kubelet[3168]: W0904 17:42:35.230294 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230312 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230534 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.230724 kubelet[3168]: W0904 17:42:35.230545 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.230724 kubelet[3168]: E0904 17:42:35.230568 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.231546 kubelet[3168]: E0904 17:42:35.230994 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.231546 kubelet[3168]: W0904 17:42:35.231006 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.231546 kubelet[3168]: E0904 17:42:35.231024 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.231546 kubelet[3168]: E0904 17:42:35.231257 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.231546 kubelet[3168]: W0904 17:42:35.231269 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.231546 kubelet[3168]: E0904 17:42:35.231391 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.231801 kubelet[3168]: E0904 17:42:35.231570 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.231801 kubelet[3168]: W0904 17:42:35.231580 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.231801 kubelet[3168]: E0904 17:42:35.231665 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.231942 kubelet[3168]: E0904 17:42:35.231849 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.231942 kubelet[3168]: W0904 17:42:35.231858 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.232029 kubelet[3168]: E0904 17:42:35.231880 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232154 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233063 kubelet[3168]: W0904 17:42:35.232167 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232203 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232416 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233063 kubelet[3168]: W0904 17:42:35.232427 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232445 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232674 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233063 kubelet[3168]: W0904 17:42:35.232686 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.232787 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233063 kubelet[3168]: E0904 17:42:35.233000 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233689 kubelet[3168]: W0904 17:42:35.233011 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233689 kubelet[3168]: E0904 17:42:35.233199 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233689 kubelet[3168]: E0904 17:42:35.233273 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233689 kubelet[3168]: W0904 17:42:35.233281 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233689 kubelet[3168]: E0904 17:42:35.233362 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.233689 kubelet[3168]: E0904 17:42:35.233542 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.233689 kubelet[3168]: W0904 17:42:35.233551 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.233689 kubelet[3168]: E0904 17:42:35.233567 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.234063 kubelet[3168]: E0904 17:42:35.233755 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.234063 kubelet[3168]: W0904 17:42:35.233765 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.234063 kubelet[3168]: E0904 17:42:35.233789 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.234189 kubelet[3168]: E0904 17:42:35.234074 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.234189 kubelet[3168]: W0904 17:42:35.234085 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.234189 kubelet[3168]: E0904 17:42:35.234110 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.234378 kubelet[3168]: E0904 17:42:35.234345 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.234378 kubelet[3168]: W0904 17:42:35.234356 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.234378 kubelet[3168]: E0904 17:42:35.234372 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.234570 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.235421 kubelet[3168]: W0904 17:42:35.234583 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.234595 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.234783 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.235421 kubelet[3168]: W0904 17:42:35.234794 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.234807 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.235068 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.235421 kubelet[3168]: W0904 17:42:35.235078 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.235090 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.235421 kubelet[3168]: E0904 17:42:35.235315 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.235859 kubelet[3168]: W0904 17:42:35.235325 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.235859 kubelet[3168]: E0904 17:42:35.235337 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.334751 kubelet[3168]: E0904 17:42:35.334632 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.335282 kubelet[3168]: W0904 17:42:35.334993 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.335282 kubelet[3168]: E0904 17:42:35.335034 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.336578 kubelet[3168]: E0904 17:42:35.336393 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.336578 kubelet[3168]: W0904 17:42:35.336411 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.336578 kubelet[3168]: E0904 17:42:35.336430 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.337019 kubelet[3168]: E0904 17:42:35.336838 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.337019 kubelet[3168]: W0904 17:42:35.336853 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.337019 kubelet[3168]: E0904 17:42:35.336870 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.338676 kubelet[3168]: E0904 17:42:35.337962 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.338676 kubelet[3168]: W0904 17:42:35.337979 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.338676 kubelet[3168]: E0904 17:42:35.337994 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.338676 kubelet[3168]: E0904 17:42:35.338516 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.338676 kubelet[3168]: W0904 17:42:35.338529 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.338676 kubelet[3168]: E0904 17:42:35.338544 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.339792 kubelet[3168]: E0904 17:42:35.339748 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.340442 kubelet[3168]: W0904 17:42:35.340279 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.340442 kubelet[3168]: E0904 17:42:35.340303 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.440830 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442252 kubelet[3168]: W0904 17:42:35.440863 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.440903 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.441194 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442252 kubelet[3168]: W0904 17:42:35.441206 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.441219 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.441454 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442252 kubelet[3168]: W0904 17:42:35.441463 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.441474 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442252 kubelet[3168]: E0904 17:42:35.441675 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442846 kubelet[3168]: W0904 17:42:35.441683 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442846 kubelet[3168]: E0904 17:42:35.441694 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442846 kubelet[3168]: E0904 17:42:35.441897 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442846 kubelet[3168]: W0904 17:42:35.441907 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442846 kubelet[3168]: E0904 17:42:35.441921 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.442846 kubelet[3168]: E0904 17:42:35.442166 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.442846 kubelet[3168]: W0904 17:42:35.442175 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.442846 kubelet[3168]: E0904 17:42:35.442187 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543240 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544039 kubelet[3168]: W0904 17:42:35.543267 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543292 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543544 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544039 kubelet[3168]: W0904 17:42:35.543556 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543570 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543788 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544039 kubelet[3168]: W0904 17:42:35.543799 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.543812 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544039 kubelet[3168]: E0904 17:42:35.544049 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544598 kubelet[3168]: W0904 17:42:35.544061 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544598 kubelet[3168]: E0904 17:42:35.544075 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544598 kubelet[3168]: E0904 17:42:35.544322 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544598 kubelet[3168]: W0904 17:42:35.544332 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544598 kubelet[3168]: E0904 17:42:35.544345 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.544598 kubelet[3168]: E0904 17:42:35.544535 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.544598 kubelet[3168]: W0904 17:42:35.544545 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.544598 kubelet[3168]: E0904 17:42:35.544556 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.640565 kubelet[3168]: E0904 17:42:35.640398 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.640565 kubelet[3168]: W0904 17:42:35.640424 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.640565 kubelet[3168]: E0904 17:42:35.640454 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.645217 kubelet[3168]: E0904 17:42:35.645104 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.645217 kubelet[3168]: W0904 17:42:35.645126 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.645217 kubelet[3168]: E0904 17:42:35.645147 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.645755 kubelet[3168]: E0904 17:42:35.645576 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.645755 kubelet[3168]: W0904 17:42:35.645587 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.645755 kubelet[3168]: E0904 17:42:35.645600 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.646110 kubelet[3168]: E0904 17:42:35.645986 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.646110 kubelet[3168]: W0904 17:42:35.645998 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.646110 kubelet[3168]: E0904 17:42:35.646011 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.646643 kubelet[3168]: E0904 17:42:35.646431 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.646643 kubelet[3168]: W0904 17:42:35.646446 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.646643 kubelet[3168]: E0904 17:42:35.646463 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.646926 kubelet[3168]: E0904 17:42:35.646844 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.646926 kubelet[3168]: W0904 17:42:35.646857 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.646926 kubelet[3168]: E0904 17:42:35.646874 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.747915 kubelet[3168]: E0904 17:42:35.747837 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.747915 kubelet[3168]: W0904 17:42:35.747862 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.747915 kubelet[3168]: E0904 17:42:35.747906 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.748519 kubelet[3168]: E0904 17:42:35.748171 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.748519 kubelet[3168]: W0904 17:42:35.748182 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.748519 kubelet[3168]: E0904 17:42:35.748198 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.748519 kubelet[3168]: E0904 17:42:35.748419 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.748519 kubelet[3168]: W0904 17:42:35.748429 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.748519 kubelet[3168]: E0904 17:42:35.748443 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.748841 kubelet[3168]: E0904 17:42:35.748644 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.748841 kubelet[3168]: W0904 17:42:35.748654 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.748841 kubelet[3168]: E0904 17:42:35.748665 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.748985 kubelet[3168]: E0904 17:42:35.748869 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.748985 kubelet[3168]: W0904 17:42:35.748878 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.748985 kubelet[3168]: E0904 17:42:35.748924 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.797550 kubelet[3168]: E0904 17:42:35.797522 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.797550 kubelet[3168]: W0904 17:42:35.797540 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.797550 kubelet[3168]: E0904 17:42:35.797559 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.804350 kubelet[3168]: E0904 17:42:35.804324 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.804350 kubelet[3168]: W0904 17:42:35.804340 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.804485 kubelet[3168]: E0904 17:42:35.804359 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.849374 kubelet[3168]: E0904 17:42:35.849341 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.849374 kubelet[3168]: W0904 17:42:35.849364 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.849631 kubelet[3168]: E0904 17:42:35.849388 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.849690 kubelet[3168]: E0904 17:42:35.849659 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.849690 kubelet[3168]: W0904 17:42:35.849670 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.849690 kubelet[3168]: E0904 17:42:35.849687 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.849962 kubelet[3168]: E0904 17:42:35.849942 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.850065 kubelet[3168]: W0904 17:42:35.849956 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.850065 kubelet[3168]: E0904 17:42:35.849980 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.929195 kubelet[3168]: E0904 17:42:35.927541 3168 projected.go:294] Couldn't get configMap calico-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Sep 4 17:42:35.929195 kubelet[3168]: E0904 17:42:35.927576 3168 projected.go:200] Error preparing data for projected volume kube-api-access-8fc4k for pod calico-system/calico-typha-675d96d689-vxm87: failed to sync configmap cache: timed out waiting for the condition Sep 4 17:42:35.929195 kubelet[3168]: E0904 17:42:35.927691 3168 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cb36bc54-36ce-4298-94ef-03febcd650bf-kube-api-access-8fc4k podName:cb36bc54-36ce-4298-94ef-03febcd650bf nodeName:}" failed. No retries permitted until 2024-09-04 17:42:36.42766498 +0000 UTC m=+21.801016422 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8fc4k" (UniqueName: "kubernetes.io/projected/cb36bc54-36ce-4298-94ef-03febcd650bf-kube-api-access-8fc4k") pod "calico-typha-675d96d689-vxm87" (UID: "cb36bc54-36ce-4298-94ef-03febcd650bf") : failed to sync configmap cache: timed out waiting for the condition Sep 4 17:42:35.950934 kubelet[3168]: E0904 17:42:35.950902 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.950934 kubelet[3168]: W0904 17:42:35.950926 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.951150 kubelet[3168]: E0904 17:42:35.950951 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.951246 kubelet[3168]: E0904 17:42:35.951225 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.951246 kubelet[3168]: W0904 17:42:35.951241 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.951353 kubelet[3168]: E0904 17:42:35.951260 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:35.951505 kubelet[3168]: E0904 17:42:35.951487 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:35.951505 kubelet[3168]: W0904 17:42:35.951501 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:35.951601 kubelet[3168]: E0904 17:42:35.951517 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.054392 kubelet[3168]: E0904 17:42:36.052911 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.054392 kubelet[3168]: W0904 17:42:36.052941 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.054392 kubelet[3168]: E0904 17:42:36.052965 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.055082 kubelet[3168]: E0904 17:42:36.054920 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.055082 kubelet[3168]: W0904 17:42:36.054941 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.055082 kubelet[3168]: E0904 17:42:36.054962 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.055602 kubelet[3168]: E0904 17:42:36.055584 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.056393 kubelet[3168]: W0904 17:42:36.056161 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.056393 kubelet[3168]: E0904 17:42:36.056200 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.056758 kubelet[3168]: E0904 17:42:36.056671 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.056970 kubelet[3168]: W0904 17:42:36.056838 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.056970 kubelet[3168]: E0904 17:42:36.056862 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.072046 kubelet[3168]: E0904 17:42:36.072023 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.073331 kubelet[3168]: W0904 17:42:36.072363 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.073331 kubelet[3168]: E0904 17:42:36.072395 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.111288 containerd[1676]: time="2024-09-04T17:42:36.111241281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rcds4,Uid:2f2ca5b5-17c8-43a2-9515-098a3c51c0f8,Namespace:calico-system,Attempt:0,}" Sep 4 17:42:36.158957 kubelet[3168]: E0904 17:42:36.157609 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.158957 kubelet[3168]: W0904 17:42:36.157636 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.158957 kubelet[3168]: E0904 17:42:36.157663 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.171834 containerd[1676]: time="2024-09-04T17:42:36.171662236Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:36.171834 containerd[1676]: time="2024-09-04T17:42:36.171718336Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:36.171834 containerd[1676]: time="2024-09-04T17:42:36.171732037Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:36.172664 containerd[1676]: time="2024-09-04T17:42:36.171820838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:36.199044 systemd[1]: Started cri-containerd-16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4.scope - libcontainer container 16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4. Sep 4 17:42:36.220801 containerd[1676]: time="2024-09-04T17:42:36.220612909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rcds4,Uid:2f2ca5b5-17c8-43a2-9515-098a3c51c0f8,Namespace:calico-system,Attempt:0,} returns sandbox id \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\"" Sep 4 17:42:36.223845 containerd[1676]: time="2024-09-04T17:42:36.223627757Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\"" Sep 4 17:42:36.258830 kubelet[3168]: E0904 17:42:36.258795 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.258830 kubelet[3168]: W0904 17:42:36.258822 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.258979 kubelet[3168]: E0904 17:42:36.258846 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.360153 kubelet[3168]: E0904 17:42:36.360110 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.360153 kubelet[3168]: W0904 17:42:36.360135 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.360339 kubelet[3168]: E0904 17:42:36.360162 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.461007 kubelet[3168]: E0904 17:42:36.460863 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.461472 kubelet[3168]: W0904 17:42:36.461192 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.461472 kubelet[3168]: E0904 17:42:36.461231 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.462005 kubelet[3168]: E0904 17:42:36.461775 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.462005 kubelet[3168]: W0904 17:42:36.461823 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.462005 kubelet[3168]: E0904 17:42:36.461842 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.462486 kubelet[3168]: E0904 17:42:36.462441 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.462657 kubelet[3168]: W0904 17:42:36.462455 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.462657 kubelet[3168]: E0904 17:42:36.462583 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.463129 kubelet[3168]: E0904 17:42:36.463108 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.463215 kubelet[3168]: W0904 17:42:36.463135 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.463215 kubelet[3168]: E0904 17:42:36.463151 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.463661 kubelet[3168]: E0904 17:42:36.463583 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.463661 kubelet[3168]: W0904 17:42:36.463600 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.463661 kubelet[3168]: E0904 17:42:36.463614 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.469731 kubelet[3168]: E0904 17:42:36.468999 3168 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:42:36.469731 kubelet[3168]: W0904 17:42:36.469015 3168 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:42:36.469731 kubelet[3168]: E0904 17:42:36.469030 3168 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:42:36.528921 containerd[1676]: time="2024-09-04T17:42:36.528637577Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-675d96d689-vxm87,Uid:cb36bc54-36ce-4298-94ef-03febcd650bf,Namespace:calico-system,Attempt:0,}" Sep 4 17:42:36.594161 containerd[1676]: time="2024-09-04T17:42:36.594077811Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:36.594376 containerd[1676]: time="2024-09-04T17:42:36.594147112Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:36.594376 containerd[1676]: time="2024-09-04T17:42:36.594191613Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:36.594376 containerd[1676]: time="2024-09-04T17:42:36.594293315Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:36.611065 systemd[1]: Started cri-containerd-80b6569fb58cc9d0f5b9e7cc61a79cc0ea091b933971f53b34af22e7223d05e6.scope - libcontainer container 80b6569fb58cc9d0f5b9e7cc61a79cc0ea091b933971f53b34af22e7223d05e6. Sep 4 17:42:36.662955 containerd[1676]: time="2024-09-04T17:42:36.662851198Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-675d96d689-vxm87,Uid:cb36bc54-36ce-4298-94ef-03febcd650bf,Namespace:calico-system,Attempt:0,} returns sandbox id \"80b6569fb58cc9d0f5b9e7cc61a79cc0ea091b933971f53b34af22e7223d05e6\"" Sep 4 17:42:36.719026 kubelet[3168]: E0904 17:42:36.717306 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:37.439946 containerd[1676]: time="2024-09-04T17:42:37.439878978Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:37.444456 containerd[1676]: time="2024-09-04T17:42:37.444382849Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1: active requests=0, bytes read=5141007" Sep 4 17:42:37.449756 containerd[1676]: time="2024-09-04T17:42:37.449697633Z" level=info msg="ImageCreate event name:\"sha256:00564b1c843430f804fda219f98769c25b538adebc11504477d5ee331fd8f85b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:37.454140 containerd[1676]: time="2024-09-04T17:42:37.454082203Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:7938ad0cb2b49a32937962cc40dd826ad5858999c603bdf5fbf2092a4d50cf01\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:37.454931 containerd[1676]: time="2024-09-04T17:42:37.454715513Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" with image id \"sha256:00564b1c843430f804fda219f98769c25b538adebc11504477d5ee331fd8f85b\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:7938ad0cb2b49a32937962cc40dd826ad5858999c603bdf5fbf2092a4d50cf01\", size \"6633368\" in 1.229839936s" Sep 4 17:42:37.454931 containerd[1676]: time="2024-09-04T17:42:37.454776114Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" returns image reference \"sha256:00564b1c843430f804fda219f98769c25b538adebc11504477d5ee331fd8f85b\"" Sep 4 17:42:37.456588 containerd[1676]: time="2024-09-04T17:42:37.456538741Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.1\"" Sep 4 17:42:37.457670 containerd[1676]: time="2024-09-04T17:42:37.457641859Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 4 17:42:37.515661 containerd[1676]: time="2024-09-04T17:42:37.515613475Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb\"" Sep 4 17:42:37.516463 containerd[1676]: time="2024-09-04T17:42:37.516280686Z" level=info msg="StartContainer for \"7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb\"" Sep 4 17:42:37.552571 systemd[1]: run-containerd-runc-k8s.io-7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb-runc.esrRsh.mount: Deactivated successfully. Sep 4 17:42:37.563036 systemd[1]: Started cri-containerd-7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb.scope - libcontainer container 7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb. Sep 4 17:42:37.603722 containerd[1676]: time="2024-09-04T17:42:37.603642966Z" level=info msg="StartContainer for \"7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb\" returns successfully" Sep 4 17:42:37.615142 systemd[1]: cri-containerd-7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb.scope: Deactivated successfully. Sep 4 17:42:38.155914 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb-rootfs.mount: Deactivated successfully. Sep 4 17:42:38.351100 containerd[1676]: time="2024-09-04T17:42:38.351020978Z" level=info msg="shim disconnected" id=7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb namespace=k8s.io Sep 4 17:42:38.351100 containerd[1676]: time="2024-09-04T17:42:38.351085879Z" level=warning msg="cleaning up after shim disconnected" id=7db5dee912afac56e561a1881d1bf70e6252dd7e68c50e25ec6ec0d5e27ca6cb namespace=k8s.io Sep 4 17:42:38.351100 containerd[1676]: time="2024-09-04T17:42:38.351101179Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:38.716941 kubelet[3168]: E0904 17:42:38.716655 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:40.091363 containerd[1676]: time="2024-09-04T17:42:40.091311781Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:40.094179 containerd[1676]: time="2024-09-04T17:42:40.094126325Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.28.1: active requests=0, bytes read=29471335" Sep 4 17:42:40.103968 containerd[1676]: time="2024-09-04T17:42:40.103880879Z" level=info msg="ImageCreate event name:\"sha256:a19ab150adede78dd36481226e260735eb3b811481c6765aec79e8da6ae78b7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:40.109050 containerd[1676]: time="2024-09-04T17:42:40.109018960Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d97114d8e1e5186f1180fc8ef5f1309e0a8bf97efce35e0a0223d057d78d95fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:40.109992 containerd[1676]: time="2024-09-04T17:42:40.109656671Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.28.1\" with image id \"sha256:a19ab150adede78dd36481226e260735eb3b811481c6765aec79e8da6ae78b7f\", repo tag \"ghcr.io/flatcar/calico/typha:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d97114d8e1e5186f1180fc8ef5f1309e0a8bf97efce35e0a0223d057d78d95fb\", size \"30963728\" in 2.653070628s" Sep 4 17:42:40.109992 containerd[1676]: time="2024-09-04T17:42:40.109694671Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.1\" returns image reference \"sha256:a19ab150adede78dd36481226e260735eb3b811481c6765aec79e8da6ae78b7f\"" Sep 4 17:42:40.111609 containerd[1676]: time="2024-09-04T17:42:40.111238896Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.1\"" Sep 4 17:42:40.132514 containerd[1676]: time="2024-09-04T17:42:40.132485331Z" level=info msg="CreateContainer within sandbox \"80b6569fb58cc9d0f5b9e7cc61a79cc0ea091b933971f53b34af22e7223d05e6\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 4 17:42:40.174448 containerd[1676]: time="2024-09-04T17:42:40.174369693Z" level=info msg="CreateContainer within sandbox \"80b6569fb58cc9d0f5b9e7cc61a79cc0ea091b933971f53b34af22e7223d05e6\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"ea4e19236861ff4c133251ea734abb7ad9e028a387a77bde9e81c2d6eaf71e53\"" Sep 4 17:42:40.175258 containerd[1676]: time="2024-09-04T17:42:40.175227707Z" level=info msg="StartContainer for \"ea4e19236861ff4c133251ea734abb7ad9e028a387a77bde9e81c2d6eaf71e53\"" Sep 4 17:42:40.209301 systemd[1]: Started cri-containerd-ea4e19236861ff4c133251ea734abb7ad9e028a387a77bde9e81c2d6eaf71e53.scope - libcontainer container ea4e19236861ff4c133251ea734abb7ad9e028a387a77bde9e81c2d6eaf71e53. Sep 4 17:42:40.262015 containerd[1676]: time="2024-09-04T17:42:40.261799675Z" level=info msg="StartContainer for \"ea4e19236861ff4c133251ea734abb7ad9e028a387a77bde9e81c2d6eaf71e53\" returns successfully" Sep 4 17:42:40.717435 kubelet[3168]: E0904 17:42:40.717031 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:40.817878 kubelet[3168]: I0904 17:42:40.817813 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-675d96d689-vxm87" podStartSLOduration=3.371516598 podStartE2EDuration="6.817791862s" podCreationTimestamp="2024-09-04 17:42:34 +0000 UTC" firstStartedPulling="2024-09-04 17:42:36.664357822 +0000 UTC m=+22.037709164" lastFinishedPulling="2024-09-04 17:42:40.110633086 +0000 UTC m=+25.483984428" observedRunningTime="2024-09-04 17:42:40.817272753 +0000 UTC m=+26.190624195" watchObservedRunningTime="2024-09-04 17:42:40.817791862 +0000 UTC m=+26.191143304" Sep 4 17:42:41.810990 kubelet[3168]: I0904 17:42:41.810950 3168 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 4 17:42:42.719913 kubelet[3168]: E0904 17:42:42.718500 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:44.719647 kubelet[3168]: E0904 17:42:44.718315 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:45.542113 containerd[1676]: time="2024-09-04T17:42:45.542065740Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:45.543941 containerd[1676]: time="2024-09-04T17:42:45.543874468Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.28.1: active requests=0, bytes read=93083736" Sep 4 17:42:45.548839 containerd[1676]: time="2024-09-04T17:42:45.548785144Z" level=info msg="ImageCreate event name:\"sha256:f6d76a1259a8c22fd1c603577ee5bb8109bc40f2b3d0536d39160a027ffe9bab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:45.554012 containerd[1676]: time="2024-09-04T17:42:45.553951624Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:1cf32b2159ec9f938e747b82b9b7c74e26e17eb220e002a6a1bd6b5b1266e1fa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:45.555021 containerd[1676]: time="2024-09-04T17:42:45.554710636Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.28.1\" with image id \"sha256:f6d76a1259a8c22fd1c603577ee5bb8109bc40f2b3d0536d39160a027ffe9bab\", repo tag \"ghcr.io/flatcar/calico/cni:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:1cf32b2159ec9f938e747b82b9b7c74e26e17eb220e002a6a1bd6b5b1266e1fa\", size \"94576137\" in 5.443430939s" Sep 4 17:42:45.555021 containerd[1676]: time="2024-09-04T17:42:45.554747636Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.1\" returns image reference \"sha256:f6d76a1259a8c22fd1c603577ee5bb8109bc40f2b3d0536d39160a027ffe9bab\"" Sep 4 17:42:45.557119 containerd[1676]: time="2024-09-04T17:42:45.557083972Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 4 17:42:45.605709 containerd[1676]: time="2024-09-04T17:42:45.605661325Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508\"" Sep 4 17:42:45.607817 containerd[1676]: time="2024-09-04T17:42:45.606233334Z" level=info msg="StartContainer for \"5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508\"" Sep 4 17:42:45.640058 systemd[1]: Started cri-containerd-5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508.scope - libcontainer container 5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508. Sep 4 17:42:45.668602 containerd[1676]: time="2024-09-04T17:42:45.668534600Z" level=info msg="StartContainer for \"5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508\" returns successfully" Sep 4 17:42:46.717389 kubelet[3168]: E0904 17:42:46.716316 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:47.029063 containerd[1676]: time="2024-09-04T17:42:47.029009587Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 17:42:47.031041 systemd[1]: cri-containerd-5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508.scope: Deactivated successfully. Sep 4 17:42:47.053366 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508-rootfs.mount: Deactivated successfully. Sep 4 17:42:47.075697 kubelet[3168]: I0904 17:42:47.075469 3168 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.093512 3168 topology_manager.go:215] "Topology Admit Handler" podUID="d122f4f6-0879-4535-8e4c-8d3004a9ecee" podNamespace="kube-system" podName="coredns-7db6d8ff4d-7z5hw" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.097713 3168 topology_manager.go:215] "Topology Admit Handler" podUID="09e4b25b-6a32-4833-a01e-38ce1d8f419d" podNamespace="kube-system" podName="coredns-7db6d8ff4d-54gwk" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.099651 3168 topology_manager.go:215] "Topology Admit Handler" podUID="fda5f1ed-c7a7-4efd-a21f-f27816673cb7" podNamespace="calico-system" podName="calico-kube-controllers-f9c88fd74-zz8sb" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.138220 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrj8f\" (UniqueName: \"kubernetes.io/projected/09e4b25b-6a32-4833-a01e-38ce1d8f419d-kube-api-access-vrj8f\") pod \"coredns-7db6d8ff4d-54gwk\" (UID: \"09e4b25b-6a32-4833-a01e-38ce1d8f419d\") " pod="kube-system/coredns-7db6d8ff4d-54gwk" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.138254 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fda5f1ed-c7a7-4efd-a21f-f27816673cb7-tigera-ca-bundle\") pod \"calico-kube-controllers-f9c88fd74-zz8sb\" (UID: \"fda5f1ed-c7a7-4efd-a21f-f27816673cb7\") " pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" Sep 4 17:42:47.527139 kubelet[3168]: I0904 17:42:47.138313 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhslz\" (UniqueName: \"kubernetes.io/projected/fda5f1ed-c7a7-4efd-a21f-f27816673cb7-kube-api-access-mhslz\") pod \"calico-kube-controllers-f9c88fd74-zz8sb\" (UID: \"fda5f1ed-c7a7-4efd-a21f-f27816673cb7\") " pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" Sep 4 17:42:47.109203 systemd[1]: Created slice kubepods-burstable-podd122f4f6_0879_4535_8e4c_8d3004a9ecee.slice - libcontainer container kubepods-burstable-podd122f4f6_0879_4535_8e4c_8d3004a9ecee.slice. Sep 4 17:42:47.527763 kubelet[3168]: I0904 17:42:47.138348 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09e4b25b-6a32-4833-a01e-38ce1d8f419d-config-volume\") pod \"coredns-7db6d8ff4d-54gwk\" (UID: \"09e4b25b-6a32-4833-a01e-38ce1d8f419d\") " pod="kube-system/coredns-7db6d8ff4d-54gwk" Sep 4 17:42:47.527763 kubelet[3168]: I0904 17:42:47.138375 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rblkd\" (UniqueName: \"kubernetes.io/projected/d122f4f6-0879-4535-8e4c-8d3004a9ecee-kube-api-access-rblkd\") pod \"coredns-7db6d8ff4d-7z5hw\" (UID: \"d122f4f6-0879-4535-8e4c-8d3004a9ecee\") " pod="kube-system/coredns-7db6d8ff4d-7z5hw" Sep 4 17:42:47.527763 kubelet[3168]: I0904 17:42:47.138394 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d122f4f6-0879-4535-8e4c-8d3004a9ecee-config-volume\") pod \"coredns-7db6d8ff4d-7z5hw\" (UID: \"d122f4f6-0879-4535-8e4c-8d3004a9ecee\") " pod="kube-system/coredns-7db6d8ff4d-7z5hw" Sep 4 17:42:47.120521 systemd[1]: Created slice kubepods-burstable-pod09e4b25b_6a32_4833_a01e_38ce1d8f419d.slice - libcontainer container kubepods-burstable-pod09e4b25b_6a32_4833_a01e_38ce1d8f419d.slice. Sep 4 17:42:47.129507 systemd[1]: Created slice kubepods-besteffort-podfda5f1ed_c7a7_4efd_a21f_f27816673cb7.slice - libcontainer container kubepods-besteffort-podfda5f1ed_c7a7_4efd_a21f_f27816673cb7.slice. Sep 4 17:42:47.830179 containerd[1676]: time="2024-09-04T17:42:47.830018103Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-7z5hw,Uid:d122f4f6-0879-4535-8e4c-8d3004a9ecee,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:47.837744 containerd[1676]: time="2024-09-04T17:42:47.837706622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-54gwk,Uid:09e4b25b-6a32-4833-a01e-38ce1d8f419d,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:47.839499 containerd[1676]: time="2024-09-04T17:42:47.839455849Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f9c88fd74-zz8sb,Uid:fda5f1ed-c7a7-4efd-a21f-f27816673cb7,Namespace:calico-system,Attempt:0,}" Sep 4 17:42:48.724316 systemd[1]: Created slice kubepods-besteffort-poda4bba15f_9dd6_4ea9_9397_c6b46c29bb6d.slice - libcontainer container kubepods-besteffort-poda4bba15f_9dd6_4ea9_9397_c6b46c29bb6d.slice. Sep 4 17:42:48.726801 containerd[1676]: time="2024-09-04T17:42:48.726680501Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-msd5w,Uid:a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d,Namespace:calico-system,Attempt:0,}" Sep 4 17:42:49.149556 containerd[1676]: time="2024-09-04T17:42:49.149490455Z" level=info msg="shim disconnected" id=5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508 namespace=k8s.io Sep 4 17:42:49.149556 containerd[1676]: time="2024-09-04T17:42:49.149551555Z" level=warning msg="cleaning up after shim disconnected" id=5f6f7700d24759bca71c1ea5162274e3338d680812ab6e41e45191e7935fe508 namespace=k8s.io Sep 4 17:42:49.149556 containerd[1676]: time="2024-09-04T17:42:49.149562356Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:49.426684 containerd[1676]: time="2024-09-04T17:42:49.426538549Z" level=error msg="Failed to destroy network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.427968 containerd[1676]: time="2024-09-04T17:42:49.427693567Z" level=error msg="encountered an error cleaning up failed sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.427968 containerd[1676]: time="2024-09-04T17:42:49.427768968Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-7z5hw,Uid:d122f4f6-0879-4535-8e4c-8d3004a9ecee,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.430558 kubelet[3168]: E0904 17:42:49.430062 3168 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.430558 kubelet[3168]: E0904 17:42:49.430160 3168 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-7z5hw" Sep 4 17:42:49.430558 kubelet[3168]: E0904 17:42:49.430188 3168 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-7z5hw" Sep 4 17:42:49.431119 kubelet[3168]: E0904 17:42:49.430242 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-7z5hw_kube-system(d122f4f6-0879-4535-8e4c-8d3004a9ecee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-7z5hw_kube-system(d122f4f6-0879-4535-8e4c-8d3004a9ecee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-7z5hw" podUID="d122f4f6-0879-4535-8e4c-8d3004a9ecee" Sep 4 17:42:49.434846 containerd[1676]: time="2024-09-04T17:42:49.434718176Z" level=error msg="Failed to destroy network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.435288 containerd[1676]: time="2024-09-04T17:42:49.435221983Z" level=error msg="encountered an error cleaning up failed sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.435437 containerd[1676]: time="2024-09-04T17:42:49.435382586Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f9c88fd74-zz8sb,Uid:fda5f1ed-c7a7-4efd-a21f-f27816673cb7,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.436120 kubelet[3168]: E0904 17:42:49.435711 3168 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.436120 kubelet[3168]: E0904 17:42:49.435764 3168 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" Sep 4 17:42:49.436120 kubelet[3168]: E0904 17:42:49.435788 3168 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" Sep 4 17:42:49.436455 kubelet[3168]: E0904 17:42:49.435833 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-f9c88fd74-zz8sb_calico-system(fda5f1ed-c7a7-4efd-a21f-f27816673cb7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-f9c88fd74-zz8sb_calico-system(fda5f1ed-c7a7-4efd-a21f-f27816673cb7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" podUID="fda5f1ed-c7a7-4efd-a21f-f27816673cb7" Sep 4 17:42:49.438381 containerd[1676]: time="2024-09-04T17:42:49.438299731Z" level=error msg="Failed to destroy network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.438841 containerd[1676]: time="2024-09-04T17:42:49.438803239Z" level=error msg="encountered an error cleaning up failed sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.439072 containerd[1676]: time="2024-09-04T17:42:49.438859940Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-54gwk,Uid:09e4b25b-6a32-4833-a01e-38ce1d8f419d,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.439384 kubelet[3168]: E0904 17:42:49.439170 3168 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.439384 kubelet[3168]: E0904 17:42:49.439222 3168 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-54gwk" Sep 4 17:42:49.439384 kubelet[3168]: E0904 17:42:49.439247 3168 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-54gwk" Sep 4 17:42:49.439542 kubelet[3168]: E0904 17:42:49.439317 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-54gwk_kube-system(09e4b25b-6a32-4833-a01e-38ce1d8f419d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-54gwk_kube-system(09e4b25b-6a32-4833-a01e-38ce1d8f419d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-54gwk" podUID="09e4b25b-6a32-4833-a01e-38ce1d8f419d" Sep 4 17:42:49.443311 containerd[1676]: time="2024-09-04T17:42:49.443280508Z" level=error msg="Failed to destroy network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.443694 containerd[1676]: time="2024-09-04T17:42:49.443665514Z" level=error msg="encountered an error cleaning up failed sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.443923 containerd[1676]: time="2024-09-04T17:42:49.443796716Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-msd5w,Uid:a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.444157 kubelet[3168]: E0904 17:42:49.444064 3168 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.444157 kubelet[3168]: E0904 17:42:49.444103 3168 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:49.444157 kubelet[3168]: E0904 17:42:49.444124 3168 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-msd5w" Sep 4 17:42:49.444568 kubelet[3168]: E0904 17:42:49.444357 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-msd5w_calico-system(a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-msd5w_calico-system(a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:49.834632 kubelet[3168]: I0904 17:42:49.833589 3168 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:42:49.834798 containerd[1676]: time="2024-09-04T17:42:49.834391770Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.1\"" Sep 4 17:42:49.836187 containerd[1676]: time="2024-09-04T17:42:49.835634590Z" level=info msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" Sep 4 17:42:49.836187 containerd[1676]: time="2024-09-04T17:42:49.835982695Z" level=info msg="Ensure that sandbox 50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84 in task-service has been cleanup successfully" Sep 4 17:42:49.840572 kubelet[3168]: I0904 17:42:49.839658 3168 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:42:49.842297 containerd[1676]: time="2024-09-04T17:42:49.841575282Z" level=info msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" Sep 4 17:42:49.842297 containerd[1676]: time="2024-09-04T17:42:49.841790885Z" level=info msg="Ensure that sandbox 2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5 in task-service has been cleanup successfully" Sep 4 17:42:49.847541 kubelet[3168]: I0904 17:42:49.846512 3168 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:42:49.848179 containerd[1676]: time="2024-09-04T17:42:49.848146284Z" level=info msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" Sep 4 17:42:49.848387 containerd[1676]: time="2024-09-04T17:42:49.848355487Z" level=info msg="Ensure that sandbox 8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849 in task-service has been cleanup successfully" Sep 4 17:42:49.855303 kubelet[3168]: I0904 17:42:49.855278 3168 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:42:49.863387 containerd[1676]: time="2024-09-04T17:42:49.863335019Z" level=info msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" Sep 4 17:42:49.863794 containerd[1676]: time="2024-09-04T17:42:49.863757026Z" level=info msg="Ensure that sandbox d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86 in task-service has been cleanup successfully" Sep 4 17:42:49.923586 containerd[1676]: time="2024-09-04T17:42:49.923325049Z" level=error msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" failed" error="failed to destroy network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.924722 kubelet[3168]: E0904 17:42:49.924489 3168 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:42:49.924722 kubelet[3168]: E0904 17:42:49.924567 3168 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5"} Sep 4 17:42:49.924722 kubelet[3168]: E0904 17:42:49.924647 3168 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:42:49.924722 kubelet[3168]: E0904 17:42:49.924680 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-msd5w" podUID="a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d" Sep 4 17:42:49.929178 containerd[1676]: time="2024-09-04T17:42:49.929132939Z" level=error msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" failed" error="failed to destroy network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.929505 kubelet[3168]: E0904 17:42:49.929362 3168 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:42:49.929505 kubelet[3168]: E0904 17:42:49.929411 3168 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84"} Sep 4 17:42:49.929505 kubelet[3168]: E0904 17:42:49.929449 3168 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"09e4b25b-6a32-4833-a01e-38ce1d8f419d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:42:49.929505 kubelet[3168]: E0904 17:42:49.929477 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"09e4b25b-6a32-4833-a01e-38ce1d8f419d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-54gwk" podUID="09e4b25b-6a32-4833-a01e-38ce1d8f419d" Sep 4 17:42:49.932508 containerd[1676]: time="2024-09-04T17:42:49.932432390Z" level=error msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" failed" error="failed to destroy network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.932689 kubelet[3168]: E0904 17:42:49.932659 3168 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:42:49.932770 kubelet[3168]: E0904 17:42:49.932699 3168 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849"} Sep 4 17:42:49.932770 kubelet[3168]: E0904 17:42:49.932733 3168 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"fda5f1ed-c7a7-4efd-a21f-f27816673cb7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:42:49.932932 kubelet[3168]: E0904 17:42:49.932764 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"fda5f1ed-c7a7-4efd-a21f-f27816673cb7\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" podUID="fda5f1ed-c7a7-4efd-a21f-f27816673cb7" Sep 4 17:42:49.934509 containerd[1676]: time="2024-09-04T17:42:49.934469722Z" level=error msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" failed" error="failed to destroy network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:42:49.934673 kubelet[3168]: E0904 17:42:49.934642 3168 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:42:49.934739 kubelet[3168]: E0904 17:42:49.934696 3168 kuberuntime_manager.go:1375] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86"} Sep 4 17:42:49.934739 kubelet[3168]: E0904 17:42:49.934729 3168 kuberuntime_manager.go:1075] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"d122f4f6-0879-4535-8e4c-8d3004a9ecee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:42:49.934850 kubelet[3168]: E0904 17:42:49.934772 3168 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"d122f4f6-0879-4535-8e4c-8d3004a9ecee\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-7z5hw" podUID="d122f4f6-0879-4535-8e4c-8d3004a9ecee" Sep 4 17:42:50.258427 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5-shm.mount: Deactivated successfully. Sep 4 17:42:50.258578 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849-shm.mount: Deactivated successfully. Sep 4 17:42:50.258685 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84-shm.mount: Deactivated successfully. Sep 4 17:42:50.258784 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86-shm.mount: Deactivated successfully. Sep 4 17:42:59.580759 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount901674277.mount: Deactivated successfully. Sep 4 17:42:59.879046 containerd[1676]: time="2024-09-04T17:42:59.878871302Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:42:59.951406 containerd[1676]: time="2024-09-04T17:42:59.951158234Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.28.1: active requests=0, bytes read=117873564" Sep 4 17:43:00.008230 containerd[1676]: time="2024-09-04T17:43:00.008136725Z" level=info msg="ImageCreate event name:\"sha256:8bbeb9e1ee3287b8f750c10383f53fa1ec6f942aaea2a900f666d5e4e63cf4cc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:00.032444 containerd[1676]: time="2024-09-04T17:43:00.032287603Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:47908d8b3046dadd6fbea273ac5b0b9bb803cc7b58b9114c50bf7591767d2744\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:00.033511 containerd[1676]: time="2024-09-04T17:43:00.033100816Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.28.1\" with image id \"sha256:8bbeb9e1ee3287b8f750c10383f53fa1ec6f942aaea2a900f666d5e4e63cf4cc\", repo tag \"ghcr.io/flatcar/calico/node:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:47908d8b3046dadd6fbea273ac5b0b9bb803cc7b58b9114c50bf7591767d2744\", size \"117873426\" in 10.198656245s" Sep 4 17:43:00.033511 containerd[1676]: time="2024-09-04T17:43:00.033150417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.1\" returns image reference \"sha256:8bbeb9e1ee3287b8f750c10383f53fa1ec6f942aaea2a900f666d5e4e63cf4cc\"" Sep 4 17:43:00.049533 containerd[1676]: time="2024-09-04T17:43:00.049493873Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 4 17:43:00.426639 containerd[1676]: time="2024-09-04T17:43:00.426580375Z" level=info msg="CreateContainer within sandbox \"16f984ed6b7042190a7f35545bf6a5e48ec0c295126cab93beeab0e67df359c4\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d\"" Sep 4 17:43:00.428101 containerd[1676]: time="2024-09-04T17:43:00.427315586Z" level=info msg="StartContainer for \"e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d\"" Sep 4 17:43:00.463044 systemd[1]: Started cri-containerd-e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d.scope - libcontainer container e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d. Sep 4 17:43:00.491719 containerd[1676]: time="2024-09-04T17:43:00.491586292Z" level=info msg="StartContainer for \"e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d\" returns successfully" Sep 4 17:43:00.763538 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 4 17:43:00.763702 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 4 17:43:00.914838 kubelet[3168]: I0904 17:43:00.914489 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-rcds4" podStartSLOduration=3.1035371290000002 podStartE2EDuration="26.914463811s" podCreationTimestamp="2024-09-04 17:42:34 +0000 UTC" firstStartedPulling="2024-09-04 17:42:36.223294752 +0000 UTC m=+21.596646094" lastFinishedPulling="2024-09-04 17:43:00.034221334 +0000 UTC m=+45.407572776" observedRunningTime="2024-09-04 17:43:00.911339062 +0000 UTC m=+46.284690504" watchObservedRunningTime="2024-09-04 17:43:00.914463811 +0000 UTC m=+46.287815253" Sep 4 17:43:01.718453 containerd[1676]: time="2024-09-04T17:43:01.717453279Z" level=info msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.761 [INFO][4223] k8s.go 608: Cleaning up netns ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.761 [INFO][4223] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" iface="eth0" netns="/var/run/netns/cni-d9658d13-7e00-c191-06ae-468288f70c83" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.762 [INFO][4223] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" iface="eth0" netns="/var/run/netns/cni-d9658d13-7e00-c191-06ae-468288f70c83" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.762 [INFO][4223] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" iface="eth0" netns="/var/run/netns/cni-d9658d13-7e00-c191-06ae-468288f70c83" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.762 [INFO][4223] k8s.go 615: Releasing IP address(es) ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.762 [INFO][4223] utils.go 188: Calico CNI releasing IP address ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.782 [INFO][4230] ipam_plugin.go 417: Releasing address using handleID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.783 [INFO][4230] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.783 [INFO][4230] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.787 [WARNING][4230] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.787 [INFO][4230] ipam_plugin.go 445: Releasing address using workloadID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.788 [INFO][4230] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:01.791873 containerd[1676]: 2024-09-04 17:43:01.790 [INFO][4223] k8s.go 621: Teardown processing complete. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:01.792470 containerd[1676]: time="2024-09-04T17:43:01.792050246Z" level=info msg="TearDown network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" successfully" Sep 4 17:43:01.792470 containerd[1676]: time="2024-09-04T17:43:01.792101747Z" level=info msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" returns successfully" Sep 4 17:43:01.795909 containerd[1676]: time="2024-09-04T17:43:01.795531601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-54gwk,Uid:09e4b25b-6a32-4833-a01e-38ce1d8f419d,Namespace:kube-system,Attempt:1,}" Sep 4 17:43:01.796458 systemd[1]: run-netns-cni\x2dd9658d13\x2d7e00\x2dc191\x2d06ae\x2d468288f70c83.mount: Deactivated successfully. Sep 4 17:43:02.517483 systemd-networkd[1471]: cali21ebde6d788: Link UP Sep 4 17:43:02.518191 systemd-networkd[1471]: cali21ebde6d788: Gained carrier Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.445 [INFO][4349] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.454 [INFO][4349] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0 coredns-7db6d8ff4d- kube-system 09e4b25b-6a32-4833-a01e-38ce1d8f419d 676 0 2024-09-04 17:42:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 coredns-7db6d8ff4d-54gwk eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali21ebde6d788 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.454 [INFO][4349] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.479 [INFO][4360] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" HandleID="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.486 [INFO][4360] ipam_plugin.go 270: Auto assigning IP ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" HandleID="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000265e60), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4054.1.0-a-b8703b3246", "pod":"coredns-7db6d8ff4d-54gwk", "timestamp":"2024-09-04 17:43:02.479705009 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.487 [INFO][4360] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.487 [INFO][4360] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.487 [INFO][4360] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.488 [INFO][4360] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.491 [INFO][4360] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.494 [INFO][4360] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.496 [INFO][4360] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.499 [INFO][4360] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.499 [INFO][4360] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.500 [INFO][4360] ipam.go 1685: Creating new handle: k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.504 [INFO][4360] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.507 [INFO][4360] ipam.go 1216: Successfully claimed IPs: [192.168.18.129/26] block=192.168.18.128/26 handle="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.507 [INFO][4360] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.129/26] handle="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.507 [INFO][4360] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:02.535316 containerd[1676]: 2024-09-04 17:43:02.507 [INFO][4360] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.129/26] IPv6=[] ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" HandleID="k8s-pod-network.6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.509 [INFO][4349] k8s.go 386: Populated endpoint ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"09e4b25b-6a32-4833-a01e-38ce1d8f419d", ResourceVersion:"676", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"coredns-7db6d8ff4d-54gwk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali21ebde6d788", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.509 [INFO][4349] k8s.go 387: Calico CNI using IPs: [192.168.18.129/32] ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.509 [INFO][4349] dataplane_linux.go 68: Setting the host side veth name to cali21ebde6d788 ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.518 [INFO][4349] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.519 [INFO][4349] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"09e4b25b-6a32-4833-a01e-38ce1d8f419d", ResourceVersion:"676", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd", Pod:"coredns-7db6d8ff4d-54gwk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali21ebde6d788", MAC:"ae:ca:db:27:3c:95", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:02.536579 containerd[1676]: 2024-09-04 17:43:02.533 [INFO][4349] k8s.go 500: Wrote updated endpoint to datastore ContainerID="6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd" Namespace="kube-system" Pod="coredns-7db6d8ff4d-54gwk" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:02.652744 containerd[1676]: time="2024-09-04T17:43:02.651967405Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:02.652744 containerd[1676]: time="2024-09-04T17:43:02.652636016Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:02.652744 containerd[1676]: time="2024-09-04T17:43:02.652658816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:02.653114 containerd[1676]: time="2024-09-04T17:43:02.652750418Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:02.676052 systemd[1]: Started cri-containerd-6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd.scope - libcontainer container 6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd. Sep 4 17:43:02.712714 containerd[1676]: time="2024-09-04T17:43:02.712667155Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-54gwk,Uid:09e4b25b-6a32-4833-a01e-38ce1d8f419d,Namespace:kube-system,Attempt:1,} returns sandbox id \"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd\"" Sep 4 17:43:02.716954 containerd[1676]: time="2024-09-04T17:43:02.716692418Z" level=info msg="CreateContainer within sandbox \"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:43:02.716954 containerd[1676]: time="2024-09-04T17:43:02.716754419Z" level=info msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.758 [INFO][4433] k8s.go 608: Cleaning up netns ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.758 [INFO][4433] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" iface="eth0" netns="/var/run/netns/cni-8d1adda9-0e45-25d6-aad1-01b303d12b40" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.758 [INFO][4433] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" iface="eth0" netns="/var/run/netns/cni-8d1adda9-0e45-25d6-aad1-01b303d12b40" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.759 [INFO][4433] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" iface="eth0" netns="/var/run/netns/cni-8d1adda9-0e45-25d6-aad1-01b303d12b40" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.759 [INFO][4433] k8s.go 615: Releasing IP address(es) ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.759 [INFO][4433] utils.go 188: Calico CNI releasing IP address ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.786 [INFO][4440] ipam_plugin.go 417: Releasing address using handleID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.786 [INFO][4440] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.786 [INFO][4440] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.791 [WARNING][4440] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.791 [INFO][4440] ipam_plugin.go 445: Releasing address using workloadID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.793 [INFO][4440] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:02.798990 containerd[1676]: 2024-09-04 17:43:02.796 [INFO][4433] k8s.go 621: Teardown processing complete. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:02.801914 containerd[1676]: time="2024-09-04T17:43:02.801817051Z" level=info msg="TearDown network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" successfully" Sep 4 17:43:02.801914 containerd[1676]: time="2024-09-04T17:43:02.801856951Z" level=info msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" returns successfully" Sep 4 17:43:02.803262 containerd[1676]: time="2024-09-04T17:43:02.803194872Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-msd5w,Uid:a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d,Namespace:calico-system,Attempt:1,}" Sep 4 17:43:02.804360 systemd[1]: run-netns-cni\x2d8d1adda9\x2d0e45\x2d25d6\x2daad1\x2d01b303d12b40.mount: Deactivated successfully. Sep 4 17:43:03.187712 containerd[1676]: time="2024-09-04T17:43:03.187557088Z" level=info msg="CreateContainer within sandbox \"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"876d4b970157457deb7e1404fba52eb5f9e565d540d866f8ae13d5989492471f\"" Sep 4 17:43:03.190153 containerd[1676]: time="2024-09-04T17:43:03.189830524Z" level=info msg="StartContainer for \"876d4b970157457deb7e1404fba52eb5f9e565d540d866f8ae13d5989492471f\"" Sep 4 17:43:03.220064 systemd[1]: Started cri-containerd-876d4b970157457deb7e1404fba52eb5f9e565d540d866f8ae13d5989492471f.scope - libcontainer container 876d4b970157457deb7e1404fba52eb5f9e565d540d866f8ae13d5989492471f. Sep 4 17:43:03.285347 containerd[1676]: time="2024-09-04T17:43:03.285293318Z" level=info msg="StartContainer for \"876d4b970157457deb7e1404fba52eb5f9e565d540d866f8ae13d5989492471f\" returns successfully" Sep 4 17:43:03.566765 systemd-networkd[1471]: caliac1dbf1319c: Link UP Sep 4 17:43:03.569260 systemd-networkd[1471]: caliac1dbf1319c: Gained carrier Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.455 [INFO][4488] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.468 [INFO][4488] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0 csi-node-driver- calico-system a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d 687 0 2024-09-04 17:42:34 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:65cb9bb8f4 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 csi-node-driver-msd5w eth0 default [] [] [kns.calico-system ksa.calico-system.default] caliac1dbf1319c [] []}} ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.468 [INFO][4488] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.519 [INFO][4504] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" HandleID="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.530 [INFO][4504] ipam_plugin.go 270: Auto assigning IP ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" HandleID="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00035ea60), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4054.1.0-a-b8703b3246", "pod":"csi-node-driver-msd5w", "timestamp":"2024-09-04 17:43:03.519177078 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.530 [INFO][4504] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.530 [INFO][4504] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.530 [INFO][4504] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.531 [INFO][4504] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.536 [INFO][4504] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.542 [INFO][4504] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.544 [INFO][4504] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.547 [INFO][4504] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.547 [INFO][4504] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.548 [INFO][4504] ipam.go 1685: Creating new handle: k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.553 [INFO][4504] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.558 [INFO][4504] ipam.go 1216: Successfully claimed IPs: [192.168.18.130/26] block=192.168.18.128/26 handle="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.558 [INFO][4504] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.130/26] handle="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.558 [INFO][4504] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:03.590567 containerd[1676]: 2024-09-04 17:43:03.558 [INFO][4504] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.130/26] IPv6=[] ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" HandleID="k8s-pod-network.1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.561 [INFO][4488] k8s.go 386: Populated endpoint ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d", ResourceVersion:"687", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65cb9bb8f4", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"csi-node-driver-msd5w", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"caliac1dbf1319c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.561 [INFO][4488] k8s.go 387: Calico CNI using IPs: [192.168.18.130/32] ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.561 [INFO][4488] dataplane_linux.go 68: Setting the host side veth name to caliac1dbf1319c ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.569 [INFO][4488] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.569 [INFO][4488] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d", ResourceVersion:"687", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65cb9bb8f4", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef", Pod:"csi-node-driver-msd5w", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"caliac1dbf1319c", MAC:"06:da:a6:24:10:57", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:03.592578 containerd[1676]: 2024-09-04 17:43:03.587 [INFO][4488] k8s.go 500: Wrote updated endpoint to datastore ContainerID="1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef" Namespace="calico-system" Pod="csi-node-driver-msd5w" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:03.644479 containerd[1676]: time="2024-09-04T17:43:03.644336337Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:03.644716 containerd[1676]: time="2024-09-04T17:43:03.644462839Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:03.644716 containerd[1676]: time="2024-09-04T17:43:03.644495440Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:03.645303 containerd[1676]: time="2024-09-04T17:43:03.645205351Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:03.665071 systemd[1]: Started cri-containerd-1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef.scope - libcontainer container 1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef. Sep 4 17:43:03.687360 containerd[1676]: time="2024-09-04T17:43:03.687315710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-msd5w,Uid:a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d,Namespace:calico-system,Attempt:1,} returns sandbox id \"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef\"" Sep 4 17:43:03.689211 containerd[1676]: time="2024-09-04T17:43:03.689168839Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.1\"" Sep 4 17:43:03.718857 containerd[1676]: time="2024-09-04T17:43:03.717872388Z" level=info msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.759 [INFO][4587] k8s.go 608: Cleaning up netns ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.759 [INFO][4587] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" iface="eth0" netns="/var/run/netns/cni-a7577782-2477-ea5a-f841-36af31cee35a" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.760 [INFO][4587] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" iface="eth0" netns="/var/run/netns/cni-a7577782-2477-ea5a-f841-36af31cee35a" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.761 [INFO][4587] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" iface="eth0" netns="/var/run/netns/cni-a7577782-2477-ea5a-f841-36af31cee35a" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.761 [INFO][4587] k8s.go 615: Releasing IP address(es) ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.761 [INFO][4587] utils.go 188: Calico CNI releasing IP address ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.782 [INFO][4593] ipam_plugin.go 417: Releasing address using handleID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.782 [INFO][4593] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.782 [INFO][4593] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.787 [WARNING][4593] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.787 [INFO][4593] ipam_plugin.go 445: Releasing address using workloadID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.788 [INFO][4593] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:03.790769 containerd[1676]: 2024-09-04 17:43:03.789 [INFO][4587] k8s.go 621: Teardown processing complete. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:03.791419 containerd[1676]: time="2024-09-04T17:43:03.790944132Z" level=info msg="TearDown network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" successfully" Sep 4 17:43:03.791419 containerd[1676]: time="2024-09-04T17:43:03.790993433Z" level=info msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" returns successfully" Sep 4 17:43:03.791840 containerd[1676]: time="2024-09-04T17:43:03.791775445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f9c88fd74-zz8sb,Uid:fda5f1ed-c7a7-4efd-a21f-f27816673cb7,Namespace:calico-system,Attempt:1,}" Sep 4 17:43:03.799377 systemd[1]: run-netns-cni\x2da7577782\x2d2477\x2dea5a\x2df841\x2d36af31cee35a.mount: Deactivated successfully. Sep 4 17:43:03.902763 kubelet[3168]: I0904 17:43:03.902215 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-54gwk" podStartSLOduration=35.902192273 podStartE2EDuration="35.902192273s" podCreationTimestamp="2024-09-04 17:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:43:03.900982654 +0000 UTC m=+49.274333996" watchObservedRunningTime="2024-09-04 17:43:03.902192273 +0000 UTC m=+49.275543615" Sep 4 17:43:04.324229 systemd-networkd[1471]: cali8515104523b: Link UP Sep 4 17:43:04.324498 systemd-networkd[1471]: cali8515104523b: Gained carrier Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.252 [INFO][4605] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.262 [INFO][4605] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0 calico-kube-controllers-f9c88fd74- calico-system fda5f1ed-c7a7-4efd-a21f-f27816673cb7 696 0 2024-09-04 17:42:35 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:f9c88fd74 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 calico-kube-controllers-f9c88fd74-zz8sb eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali8515104523b [] []}} ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.262 [INFO][4605] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.291 [INFO][4615] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" HandleID="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.298 [INFO][4615] ipam_plugin.go 270: Auto assigning IP ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" HandleID="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000265e60), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4054.1.0-a-b8703b3246", "pod":"calico-kube-controllers-f9c88fd74-zz8sb", "timestamp":"2024-09-04 17:43:04.29111416 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.298 [INFO][4615] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.298 [INFO][4615] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.299 [INFO][4615] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.300 [INFO][4615] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.304 [INFO][4615] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.308 [INFO][4615] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.309 [INFO][4615] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.311 [INFO][4615] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.311 [INFO][4615] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.312 [INFO][4615] ipam.go 1685: Creating new handle: k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.315 [INFO][4615] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.319 [INFO][4615] ipam.go 1216: Successfully claimed IPs: [192.168.18.131/26] block=192.168.18.128/26 handle="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.319 [INFO][4615] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.131/26] handle="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.319 [INFO][4615] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:04.340591 containerd[1676]: 2024-09-04 17:43:04.319 [INFO][4615] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.131/26] IPv6=[] ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" HandleID="k8s-pod-network.ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.320 [INFO][4605] k8s.go 386: Populated endpoint ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0", GenerateName:"calico-kube-controllers-f9c88fd74-", Namespace:"calico-system", SelfLink:"", UID:"fda5f1ed-c7a7-4efd-a21f-f27816673cb7", ResourceVersion:"696", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f9c88fd74", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"calico-kube-controllers-f9c88fd74-zz8sb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8515104523b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.320 [INFO][4605] k8s.go 387: Calico CNI using IPs: [192.168.18.131/32] ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.320 [INFO][4605] dataplane_linux.go 68: Setting the host side veth name to cali8515104523b ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.324 [INFO][4605] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.324 [INFO][4605] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0", GenerateName:"calico-kube-controllers-f9c88fd74-", Namespace:"calico-system", SelfLink:"", UID:"fda5f1ed-c7a7-4efd-a21f-f27816673cb7", ResourceVersion:"696", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f9c88fd74", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea", Pod:"calico-kube-controllers-f9c88fd74-zz8sb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8515104523b", MAC:"a6:ed:dd:db:00:88", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:04.343102 containerd[1676]: 2024-09-04 17:43:04.338 [INFO][4605] k8s.go 500: Wrote updated endpoint to datastore ContainerID="ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea" Namespace="calico-system" Pod="calico-kube-controllers-f9c88fd74-zz8sb" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:04.434020 systemd-networkd[1471]: cali21ebde6d788: Gained IPv6LL Sep 4 17:43:04.452346 containerd[1676]: time="2024-09-04T17:43:04.452180981Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:04.452346 containerd[1676]: time="2024-09-04T17:43:04.452274183Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:04.452575 containerd[1676]: time="2024-09-04T17:43:04.452377984Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:04.452575 containerd[1676]: time="2024-09-04T17:43:04.452504086Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:04.477048 systemd[1]: Started cri-containerd-ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea.scope - libcontainer container ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea. Sep 4 17:43:04.515684 containerd[1676]: time="2024-09-04T17:43:04.515635674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f9c88fd74-zz8sb,Uid:fda5f1ed-c7a7-4efd-a21f-f27816673cb7,Namespace:calico-system,Attempt:1,} returns sandbox id \"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea\"" Sep 4 17:43:04.719139 containerd[1676]: time="2024-09-04T17:43:04.718927256Z" level=info msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.762 [INFO][4708] k8s.go 608: Cleaning up netns ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.762 [INFO][4708] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" iface="eth0" netns="/var/run/netns/cni-52cd2cf2-742e-c205-aaeb-58556134ba17" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.762 [INFO][4708] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" iface="eth0" netns="/var/run/netns/cni-52cd2cf2-742e-c205-aaeb-58556134ba17" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.762 [INFO][4708] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" iface="eth0" netns="/var/run/netns/cni-52cd2cf2-742e-c205-aaeb-58556134ba17" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.762 [INFO][4708] k8s.go 615: Releasing IP address(es) ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.763 [INFO][4708] utils.go 188: Calico CNI releasing IP address ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.781 [INFO][4714] ipam_plugin.go 417: Releasing address using handleID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.782 [INFO][4714] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.782 [INFO][4714] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.789 [WARNING][4714] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.789 [INFO][4714] ipam_plugin.go 445: Releasing address using workloadID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.791 [INFO][4714] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:04.793139 containerd[1676]: 2024-09-04 17:43:04.792 [INFO][4708] k8s.go 621: Teardown processing complete. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:04.796487 containerd[1676]: time="2024-09-04T17:43:04.793270620Z" level=info msg="TearDown network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" successfully" Sep 4 17:43:04.796487 containerd[1676]: time="2024-09-04T17:43:04.794952846Z" level=info msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" returns successfully" Sep 4 17:43:04.796487 containerd[1676]: time="2024-09-04T17:43:04.796152365Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-7z5hw,Uid:d122f4f6-0879-4535-8e4c-8d3004a9ecee,Namespace:kube-system,Attempt:1,}" Sep 4 17:43:04.797773 systemd[1]: run-netns-cni\x2d52cd2cf2\x2d742e\x2dc205\x2daaeb\x2d58556134ba17.mount: Deactivated successfully. Sep 4 17:43:05.138030 systemd-networkd[1471]: caliac1dbf1319c: Gained IPv6LL Sep 4 17:43:05.458107 systemd-networkd[1471]: cali8515104523b: Gained IPv6LL Sep 4 17:43:05.982151 systemd-networkd[1471]: cali15ff3d99061: Link UP Sep 4 17:43:05.982648 systemd-networkd[1471]: cali15ff3d99061: Gained carrier Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.901 [INFO][4743] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.912 [INFO][4743] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0 coredns-7db6d8ff4d- kube-system d122f4f6-0879-4535-8e4c-8d3004a9ecee 713 0 2024-09-04 17:42:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 coredns-7db6d8ff4d-7z5hw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali15ff3d99061 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.912 [INFO][4743] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.940 [INFO][4755] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" HandleID="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.949 [INFO][4755] ipam_plugin.go 270: Auto assigning IP ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" HandleID="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000318e60), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4054.1.0-a-b8703b3246", "pod":"coredns-7db6d8ff4d-7z5hw", "timestamp":"2024-09-04 17:43:05.940838081 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.949 [INFO][4755] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.949 [INFO][4755] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.949 [INFO][4755] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.951 [INFO][4755] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.955 [INFO][4755] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.960 [INFO][4755] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.963 [INFO][4755] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.965 [INFO][4755] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.965 [INFO][4755] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.967 [INFO][4755] ipam.go 1685: Creating new handle: k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9 Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.970 [INFO][4755] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.977 [INFO][4755] ipam.go 1216: Successfully claimed IPs: [192.168.18.132/26] block=192.168.18.128/26 handle="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.977 [INFO][4755] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.132/26] handle="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.977 [INFO][4755] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:05.998757 containerd[1676]: 2024-09-04 17:43:05.977 [INFO][4755] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.132/26] IPv6=[] ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" HandleID="k8s-pod-network.3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.978 [INFO][4743] k8s.go 386: Populated endpoint ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d122f4f6-0879-4535-8e4c-8d3004a9ecee", ResourceVersion:"713", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"coredns-7db6d8ff4d-7z5hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali15ff3d99061", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.979 [INFO][4743] k8s.go 387: Calico CNI using IPs: [192.168.18.132/32] ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.979 [INFO][4743] dataplane_linux.go 68: Setting the host side veth name to cali15ff3d99061 ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.982 [INFO][4743] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.983 [INFO][4743] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d122f4f6-0879-4535-8e4c-8d3004a9ecee", ResourceVersion:"713", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9", Pod:"coredns-7db6d8ff4d-7z5hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali15ff3d99061", MAC:"9e:ea:80:d6:6e:2a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:06.002051 containerd[1676]: 2024-09-04 17:43:05.996 [INFO][4743] k8s.go 500: Wrote updated endpoint to datastore ContainerID="3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9" Namespace="kube-system" Pod="coredns-7db6d8ff4d-7z5hw" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:06.034379 containerd[1676]: time="2024-09-04T17:43:06.034113541Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:06.034379 containerd[1676]: time="2024-09-04T17:43:06.034172242Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:06.034379 containerd[1676]: time="2024-09-04T17:43:06.034191042Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:06.034379 containerd[1676]: time="2024-09-04T17:43:06.034281544Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:06.073074 systemd[1]: Started cri-containerd-3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9.scope - libcontainer container 3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9. Sep 4 17:43:06.121614 containerd[1676]: time="2024-09-04T17:43:06.121344706Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-7z5hw,Uid:d122f4f6-0879-4535-8e4c-8d3004a9ecee,Namespace:kube-system,Attempt:1,} returns sandbox id \"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9\"" Sep 4 17:43:06.125948 containerd[1676]: time="2024-09-04T17:43:06.125750375Z" level=info msg="CreateContainer within sandbox \"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:43:06.186332 containerd[1676]: time="2024-09-04T17:43:06.186287623Z" level=info msg="CreateContainer within sandbox \"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"918d113fa493d0f6f11a8940dc2ba501140657c32574b9072d75712b4f782035\"" Sep 4 17:43:06.187161 containerd[1676]: time="2024-09-04T17:43:06.187132636Z" level=info msg="StartContainer for \"918d113fa493d0f6f11a8940dc2ba501140657c32574b9072d75712b4f782035\"" Sep 4 17:43:06.222085 systemd[1]: Started cri-containerd-918d113fa493d0f6f11a8940dc2ba501140657c32574b9072d75712b4f782035.scope - libcontainer container 918d113fa493d0f6f11a8940dc2ba501140657c32574b9072d75712b4f782035. Sep 4 17:43:06.283461 containerd[1676]: time="2024-09-04T17:43:06.283242240Z" level=info msg="StartContainer for \"918d113fa493d0f6f11a8940dc2ba501140657c32574b9072d75712b4f782035\" returns successfully" Sep 4 17:43:06.441362 kubelet[3168]: I0904 17:43:06.441107 3168 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 4 17:43:06.463564 containerd[1676]: time="2024-09-04T17:43:06.461788235Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:06.465164 containerd[1676]: time="2024-09-04T17:43:06.465114787Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.28.1: active requests=0, bytes read=7642081" Sep 4 17:43:06.471702 containerd[1676]: time="2024-09-04T17:43:06.471439886Z" level=info msg="ImageCreate event name:\"sha256:d0c7782dfd1af19483b1da01b3d6692a92c2a570a3c8c6059128fda84c838a61\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:06.477756 containerd[1676]: time="2024-09-04T17:43:06.477720784Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:01e16d03dd0c29a8e1e302455eb15c2d0326c49cbaca4bbe8dc0e2d5308c5add\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:06.478743 containerd[1676]: time="2024-09-04T17:43:06.478450495Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.28.1\" with image id \"sha256:d0c7782dfd1af19483b1da01b3d6692a92c2a570a3c8c6059128fda84c838a61\", repo tag \"ghcr.io/flatcar/calico/csi:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:01e16d03dd0c29a8e1e302455eb15c2d0326c49cbaca4bbe8dc0e2d5308c5add\", size \"9134482\" in 2.789230055s" Sep 4 17:43:06.478743 containerd[1676]: time="2024-09-04T17:43:06.478492896Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.1\" returns image reference \"sha256:d0c7782dfd1af19483b1da01b3d6692a92c2a570a3c8c6059128fda84c838a61\"" Sep 4 17:43:06.481508 containerd[1676]: time="2024-09-04T17:43:06.481327341Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\"" Sep 4 17:43:06.483835 containerd[1676]: time="2024-09-04T17:43:06.483417673Z" level=info msg="CreateContainer within sandbox \"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 4 17:43:06.526080 containerd[1676]: time="2024-09-04T17:43:06.526037740Z" level=info msg="CreateContainer within sandbox \"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"eb9eefe8ed4c48bdee0c8c5a5e30c74202d1efa2f85987d3ffc4331ae032f736\"" Sep 4 17:43:06.528018 containerd[1676]: time="2024-09-04T17:43:06.526998255Z" level=info msg="StartContainer for \"eb9eefe8ed4c48bdee0c8c5a5e30c74202d1efa2f85987d3ffc4331ae032f736\"" Sep 4 17:43:06.566073 systemd[1]: Started cri-containerd-eb9eefe8ed4c48bdee0c8c5a5e30c74202d1efa2f85987d3ffc4331ae032f736.scope - libcontainer container eb9eefe8ed4c48bdee0c8c5a5e30c74202d1efa2f85987d3ffc4331ae032f736. Sep 4 17:43:06.618334 containerd[1676]: time="2024-09-04T17:43:06.618289984Z" level=info msg="StartContainer for \"eb9eefe8ed4c48bdee0c8c5a5e30c74202d1efa2f85987d3ffc4331ae032f736\" returns successfully" Sep 4 17:43:06.703934 kernel: bpftool[4917]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Sep 4 17:43:06.954881 kubelet[3168]: I0904 17:43:06.954744 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-7z5hw" podStartSLOduration=38.95472195 podStartE2EDuration="38.95472195s" podCreationTimestamp="2024-09-04 17:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:43:06.925069786 +0000 UTC m=+52.298421128" watchObservedRunningTime="2024-09-04 17:43:06.95472195 +0000 UTC m=+52.328073292" Sep 4 17:43:07.311416 systemd-networkd[1471]: vxlan.calico: Link UP Sep 4 17:43:07.311427 systemd-networkd[1471]: vxlan.calico: Gained carrier Sep 4 17:43:08.018069 systemd-networkd[1471]: cali15ff3d99061: Gained IPv6LL Sep 4 17:43:08.885182 containerd[1676]: time="2024-09-04T17:43:08.885138689Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:08.890690 containerd[1676]: time="2024-09-04T17:43:08.890634671Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.28.1: active requests=0, bytes read=33507125" Sep 4 17:43:08.894224 containerd[1676]: time="2024-09-04T17:43:08.894153924Z" level=info msg="ImageCreate event name:\"sha256:9d19dff735fa0889ad6e741790dd1ff35dc4443f14c95bd61459ff0b9162252e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:08.898757 containerd[1676]: time="2024-09-04T17:43:08.898687992Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:9a7338f7187d4d2352fe49eedee44b191ac92557a2e71aa3de3527ed85c1641b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:08.899862 containerd[1676]: time="2024-09-04T17:43:08.899342502Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" with image id \"sha256:9d19dff735fa0889ad6e741790dd1ff35dc4443f14c95bd61459ff0b9162252e\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:9a7338f7187d4d2352fe49eedee44b191ac92557a2e71aa3de3527ed85c1641b\", size \"34999494\" in 2.417975061s" Sep 4 17:43:08.899862 containerd[1676]: time="2024-09-04T17:43:08.899384003Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" returns image reference \"sha256:9d19dff735fa0889ad6e741790dd1ff35dc4443f14c95bd61459ff0b9162252e\"" Sep 4 17:43:08.900739 containerd[1676]: time="2024-09-04T17:43:08.900706823Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\"" Sep 4 17:43:08.922901 containerd[1676]: time="2024-09-04T17:43:08.922387749Z" level=info msg="CreateContainer within sandbox \"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 4 17:43:08.982388 containerd[1676]: time="2024-09-04T17:43:08.982341749Z" level=info msg="CreateContainer within sandbox \"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e\"" Sep 4 17:43:08.983075 containerd[1676]: time="2024-09-04T17:43:08.982915458Z" level=info msg="StartContainer for \"2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e\"" Sep 4 17:43:09.021097 systemd[1]: Started cri-containerd-2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e.scope - libcontainer container 2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e. Sep 4 17:43:09.073595 containerd[1676]: time="2024-09-04T17:43:09.073543520Z" level=info msg="StartContainer for \"2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e\" returns successfully" Sep 4 17:43:09.298049 systemd-networkd[1471]: vxlan.calico: Gained IPv6LL Sep 4 17:43:09.942761 kubelet[3168]: I0904 17:43:09.942639 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-f9c88fd74-zz8sb" podStartSLOduration=30.559359759 podStartE2EDuration="34.942614179s" podCreationTimestamp="2024-09-04 17:42:35 +0000 UTC" firstStartedPulling="2024-09-04 17:43:04.516991796 +0000 UTC m=+49.890343138" lastFinishedPulling="2024-09-04 17:43:08.900246216 +0000 UTC m=+54.273597558" observedRunningTime="2024-09-04 17:43:09.939611834 +0000 UTC m=+55.312963276" watchObservedRunningTime="2024-09-04 17:43:09.942614179 +0000 UTC m=+55.315965521" Sep 4 17:43:09.968242 systemd[1]: run-containerd-runc-k8s.io-2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e-runc.T5vlPc.mount: Deactivated successfully. Sep 4 17:43:10.449918 containerd[1676]: time="2024-09-04T17:43:10.449849001Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:10.453008 containerd[1676]: time="2024-09-04T17:43:10.452953048Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1: active requests=0, bytes read=12907822" Sep 4 17:43:10.458941 containerd[1676]: time="2024-09-04T17:43:10.458904337Z" level=info msg="ImageCreate event name:\"sha256:d1ca8f023879d2e9a9a7c98dbb3252886c5b7676be9529ddb5200aa2789b233e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:10.463565 containerd[1676]: time="2024-09-04T17:43:10.463495506Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:682cc97e4580d25b7314032c008a552bb05182fac34eba82cc389113c7767076\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:10.464331 containerd[1676]: time="2024-09-04T17:43:10.464136316Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" with image id \"sha256:d1ca8f023879d2e9a9a7c98dbb3252886c5b7676be9529ddb5200aa2789b233e\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:682cc97e4580d25b7314032c008a552bb05182fac34eba82cc389113c7767076\", size \"14400175\" in 1.563384292s" Sep 4 17:43:10.464331 containerd[1676]: time="2024-09-04T17:43:10.464177216Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" returns image reference \"sha256:d1ca8f023879d2e9a9a7c98dbb3252886c5b7676be9529ddb5200aa2789b233e\"" Sep 4 17:43:10.467397 containerd[1676]: time="2024-09-04T17:43:10.467254062Z" level=info msg="CreateContainer within sandbox \"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 4 17:43:10.501046 containerd[1676]: time="2024-09-04T17:43:10.501002770Z" level=info msg="CreateContainer within sandbox \"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"a161d29a3f730334b819fdf7810d0e549f14452b19cdb0735c21d604e32ec074\"" Sep 4 17:43:10.501577 containerd[1676]: time="2024-09-04T17:43:10.501481777Z" level=info msg="StartContainer for \"a161d29a3f730334b819fdf7810d0e549f14452b19cdb0735c21d604e32ec074\"" Sep 4 17:43:10.533018 systemd[1]: Started cri-containerd-a161d29a3f730334b819fdf7810d0e549f14452b19cdb0735c21d604e32ec074.scope - libcontainer container a161d29a3f730334b819fdf7810d0e549f14452b19cdb0735c21d604e32ec074. Sep 4 17:43:10.564279 containerd[1676]: time="2024-09-04T17:43:10.563637411Z" level=info msg="StartContainer for \"a161d29a3f730334b819fdf7810d0e549f14452b19cdb0735c21d604e32ec074\" returns successfully" Sep 4 17:43:10.813291 kubelet[3168]: I0904 17:43:10.813234 3168 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 4 17:43:10.813291 kubelet[3168]: I0904 17:43:10.813270 3168 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 4 17:43:10.939329 kubelet[3168]: I0904 17:43:10.939060 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-msd5w" podStartSLOduration=30.162524852 podStartE2EDuration="36.939036952s" podCreationTimestamp="2024-09-04 17:42:34 +0000 UTC" firstStartedPulling="2024-09-04 17:43:03.688671931 +0000 UTC m=+49.062023273" lastFinishedPulling="2024-09-04 17:43:10.465184031 +0000 UTC m=+55.838535373" observedRunningTime="2024-09-04 17:43:10.938813548 +0000 UTC m=+56.312164990" watchObservedRunningTime="2024-09-04 17:43:10.939036952 +0000 UTC m=+56.312388394" Sep 4 17:43:14.701939 containerd[1676]: time="2024-09-04T17:43:14.701858894Z" level=info msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.738 [WARNING][5150] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0", GenerateName:"calico-kube-controllers-f9c88fd74-", Namespace:"calico-system", SelfLink:"", UID:"fda5f1ed-c7a7-4efd-a21f-f27816673cb7", ResourceVersion:"756", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f9c88fd74", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea", Pod:"calico-kube-controllers-f9c88fd74-zz8sb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8515104523b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.738 [INFO][5150] k8s.go 608: Cleaning up netns ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.738 [INFO][5150] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" iface="eth0" netns="" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.738 [INFO][5150] k8s.go 615: Releasing IP address(es) ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.738 [INFO][5150] utils.go 188: Calico CNI releasing IP address ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.757 [INFO][5158] ipam_plugin.go 417: Releasing address using handleID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.757 [INFO][5158] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.757 [INFO][5158] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.762 [WARNING][5158] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.762 [INFO][5158] ipam_plugin.go 445: Releasing address using workloadID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.764 [INFO][5158] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:14.766017 containerd[1676]: 2024-09-04 17:43:14.764 [INFO][5150] k8s.go 621: Teardown processing complete. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.766017 containerd[1676]: time="2024-09-04T17:43:14.765993857Z" level=info msg="TearDown network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" successfully" Sep 4 17:43:14.766017 containerd[1676]: time="2024-09-04T17:43:14.766025558Z" level=info msg="StopPodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" returns successfully" Sep 4 17:43:14.768510 containerd[1676]: time="2024-09-04T17:43:14.767251676Z" level=info msg="RemovePodSandbox for \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" Sep 4 17:43:14.768510 containerd[1676]: time="2024-09-04T17:43:14.767294877Z" level=info msg="Forcibly stopping sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\"" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.804 [WARNING][5176] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0", GenerateName:"calico-kube-controllers-f9c88fd74-", Namespace:"calico-system", SelfLink:"", UID:"fda5f1ed-c7a7-4efd-a21f-f27816673cb7", ResourceVersion:"756", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f9c88fd74", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"ccfb1b04436307b668c24f3f49c91956e5fdca426eb6e2c513451e9c978c3cea", Pod:"calico-kube-controllers-f9c88fd74-zz8sb", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.18.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8515104523b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.805 [INFO][5176] k8s.go 608: Cleaning up netns ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.805 [INFO][5176] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" iface="eth0" netns="" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.805 [INFO][5176] k8s.go 615: Releasing IP address(es) ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.805 [INFO][5176] utils.go 188: Calico CNI releasing IP address ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.823 [INFO][5182] ipam_plugin.go 417: Releasing address using handleID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.823 [INFO][5182] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.823 [INFO][5182] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.828 [WARNING][5182] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.828 [INFO][5182] ipam_plugin.go 445: Releasing address using workloadID ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" HandleID="k8s-pod-network.8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--kube--controllers--f9c88fd74--zz8sb-eth0" Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.830 [INFO][5182] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:14.832399 containerd[1676]: 2024-09-04 17:43:14.831 [INFO][5176] k8s.go 621: Teardown processing complete. ContainerID="8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849" Sep 4 17:43:14.833159 containerd[1676]: time="2024-09-04T17:43:14.832521957Z" level=info msg="TearDown network for sandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" successfully" Sep 4 17:43:14.842276 containerd[1676]: time="2024-09-04T17:43:14.842234803Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:43:14.842400 containerd[1676]: time="2024-09-04T17:43:14.842307504Z" level=info msg="RemovePodSandbox \"8bea74e190afa0085b00cd170a730780ad5c8419fff20b6488c8329f1847b849\" returns successfully" Sep 4 17:43:14.842949 containerd[1676]: time="2024-09-04T17:43:14.842917013Z" level=info msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.873 [WARNING][5200] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d122f4f6-0879-4535-8e4c-8d3004a9ecee", ResourceVersion:"737", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9", Pod:"coredns-7db6d8ff4d-7z5hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali15ff3d99061", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.873 [INFO][5200] k8s.go 608: Cleaning up netns ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.873 [INFO][5200] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" iface="eth0" netns="" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.874 [INFO][5200] k8s.go 615: Releasing IP address(es) ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.874 [INFO][5200] utils.go 188: Calico CNI releasing IP address ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.892 [INFO][5206] ipam_plugin.go 417: Releasing address using handleID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.892 [INFO][5206] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.892 [INFO][5206] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.897 [WARNING][5206] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.897 [INFO][5206] ipam_plugin.go 445: Releasing address using workloadID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.898 [INFO][5206] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:14.900828 containerd[1676]: 2024-09-04 17:43:14.899 [INFO][5200] k8s.go 621: Teardown processing complete. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.901684 containerd[1676]: time="2024-09-04T17:43:14.900870784Z" level=info msg="TearDown network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" successfully" Sep 4 17:43:14.901684 containerd[1676]: time="2024-09-04T17:43:14.900914885Z" level=info msg="StopPodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" returns successfully" Sep 4 17:43:14.901684 containerd[1676]: time="2024-09-04T17:43:14.901458493Z" level=info msg="RemovePodSandbox for \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" Sep 4 17:43:14.901684 containerd[1676]: time="2024-09-04T17:43:14.901492793Z" level=info msg="Forcibly stopping sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\"" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.932 [WARNING][5224] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d122f4f6-0879-4535-8e4c-8d3004a9ecee", ResourceVersion:"737", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"3e6fe3d3ef769fecc05a28f0dd9588c93d90993714b6f111487027bdc47092b9", Pod:"coredns-7db6d8ff4d-7z5hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali15ff3d99061", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.933 [INFO][5224] k8s.go 608: Cleaning up netns ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.933 [INFO][5224] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" iface="eth0" netns="" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.933 [INFO][5224] k8s.go 615: Releasing IP address(es) ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.933 [INFO][5224] utils.go 188: Calico CNI releasing IP address ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.954 [INFO][5231] ipam_plugin.go 417: Releasing address using handleID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.954 [INFO][5231] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.954 [INFO][5231] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.959 [WARNING][5231] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.959 [INFO][5231] ipam_plugin.go 445: Releasing address using workloadID ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" HandleID="k8s-pod-network.d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--7z5hw-eth0" Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.960 [INFO][5231] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:14.962930 containerd[1676]: 2024-09-04 17:43:14.961 [INFO][5224] k8s.go 621: Teardown processing complete. ContainerID="d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86" Sep 4 17:43:14.962930 containerd[1676]: time="2024-09-04T17:43:14.962411409Z" level=info msg="TearDown network for sandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" successfully" Sep 4 17:43:14.972159 containerd[1676]: time="2024-09-04T17:43:14.972121155Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:43:14.972275 containerd[1676]: time="2024-09-04T17:43:14.972187456Z" level=info msg="RemovePodSandbox \"d95467e1cd64c9b20063a59143ee8297f912cf629bfc73e125f64d2786987e86\" returns successfully" Sep 4 17:43:14.972719 containerd[1676]: time="2024-09-04T17:43:14.972689963Z" level=info msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.002 [WARNING][5249] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"09e4b25b-6a32-4833-a01e-38ce1d8f419d", ResourceVersion:"700", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd", Pod:"coredns-7db6d8ff4d-54gwk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali21ebde6d788", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.002 [INFO][5249] k8s.go 608: Cleaning up netns ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.002 [INFO][5249] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" iface="eth0" netns="" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.002 [INFO][5249] k8s.go 615: Releasing IP address(es) ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.003 [INFO][5249] utils.go 188: Calico CNI releasing IP address ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.021 [INFO][5255] ipam_plugin.go 417: Releasing address using handleID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.021 [INFO][5255] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.021 [INFO][5255] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.025 [WARNING][5255] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.025 [INFO][5255] ipam_plugin.go 445: Releasing address using workloadID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.026 [INFO][5255] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:15.028741 containerd[1676]: 2024-09-04 17:43:15.027 [INFO][5249] k8s.go 621: Teardown processing complete. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.028741 containerd[1676]: time="2024-09-04T17:43:15.028694105Z" level=info msg="TearDown network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" successfully" Sep 4 17:43:15.028741 containerd[1676]: time="2024-09-04T17:43:15.028724805Z" level=info msg="StopPodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" returns successfully" Sep 4 17:43:15.029795 containerd[1676]: time="2024-09-04T17:43:15.029265213Z" level=info msg="RemovePodSandbox for \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" Sep 4 17:43:15.029795 containerd[1676]: time="2024-09-04T17:43:15.029298414Z" level=info msg="Forcibly stopping sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\"" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.059 [WARNING][5273] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"09e4b25b-6a32-4833-a01e-38ce1d8f419d", ResourceVersion:"700", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"6bf933024b4fd7ef40a74d1730d956f6063669dd5ce8f81e9b9d3b805a87dedd", Pod:"coredns-7db6d8ff4d-54gwk", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.18.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali21ebde6d788", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.059 [INFO][5273] k8s.go 608: Cleaning up netns ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.060 [INFO][5273] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" iface="eth0" netns="" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.060 [INFO][5273] k8s.go 615: Releasing IP address(es) ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.060 [INFO][5273] utils.go 188: Calico CNI releasing IP address ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.078 [INFO][5279] ipam_plugin.go 417: Releasing address using handleID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.078 [INFO][5279] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.078 [INFO][5279] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.083 [WARNING][5279] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.083 [INFO][5279] ipam_plugin.go 445: Releasing address using workloadID ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" HandleID="k8s-pod-network.50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Workload="ci--4054.1.0--a--b8703b3246-k8s-coredns--7db6d8ff4d--54gwk-eth0" Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.084 [INFO][5279] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:15.086064 containerd[1676]: 2024-09-04 17:43:15.085 [INFO][5273] k8s.go 621: Teardown processing complete. ContainerID="50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84" Sep 4 17:43:15.086742 containerd[1676]: time="2024-09-04T17:43:15.086109368Z" level=info msg="TearDown network for sandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" successfully" Sep 4 17:43:15.099855 containerd[1676]: time="2024-09-04T17:43:15.099797873Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:43:15.100034 containerd[1676]: time="2024-09-04T17:43:15.099875574Z" level=info msg="RemovePodSandbox \"50d274c09040baa13445e8a1a315e0618a13f4f264f885a36ff3f515da6ead84\" returns successfully" Sep 4 17:43:15.100532 containerd[1676]: time="2024-09-04T17:43:15.100499084Z" level=info msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.131 [WARNING][5297] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65cb9bb8f4", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef", Pod:"csi-node-driver-msd5w", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"caliac1dbf1319c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.131 [INFO][5297] k8s.go 608: Cleaning up netns ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.131 [INFO][5297] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" iface="eth0" netns="" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.131 [INFO][5297] k8s.go 615: Releasing IP address(es) ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.131 [INFO][5297] utils.go 188: Calico CNI releasing IP address ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.150 [INFO][5303] ipam_plugin.go 417: Releasing address using handleID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.150 [INFO][5303] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.150 [INFO][5303] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.154 [WARNING][5303] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.154 [INFO][5303] ipam_plugin.go 445: Releasing address using workloadID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.156 [INFO][5303] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:15.158069 containerd[1676]: 2024-09-04 17:43:15.157 [INFO][5297] k8s.go 621: Teardown processing complete. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.159020 containerd[1676]: time="2024-09-04T17:43:15.158081449Z" level=info msg="TearDown network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" successfully" Sep 4 17:43:15.159020 containerd[1676]: time="2024-09-04T17:43:15.158116950Z" level=info msg="StopPodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" returns successfully" Sep 4 17:43:15.159020 containerd[1676]: time="2024-09-04T17:43:15.158717659Z" level=info msg="RemovePodSandbox for \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" Sep 4 17:43:15.159020 containerd[1676]: time="2024-09-04T17:43:15.158754959Z" level=info msg="Forcibly stopping sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\"" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.189 [WARNING][5321] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a4bba15f-9dd6-4ea9-9397-c6b46c29bb6d", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 42, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65cb9bb8f4", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"1acecf25dfc61aa0f0e717dd9e95c9554cde7dbd316052e005935853a5f1ecef", Pod:"csi-node-driver-msd5w", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.18.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"caliac1dbf1319c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.189 [INFO][5321] k8s.go 608: Cleaning up netns ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.189 [INFO][5321] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" iface="eth0" netns="" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.189 [INFO][5321] k8s.go 615: Releasing IP address(es) ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.189 [INFO][5321] utils.go 188: Calico CNI releasing IP address ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.207 [INFO][5327] ipam_plugin.go 417: Releasing address using handleID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.207 [INFO][5327] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.207 [INFO][5327] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.213 [WARNING][5327] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.214 [INFO][5327] ipam_plugin.go 445: Releasing address using workloadID ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" HandleID="k8s-pod-network.2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Workload="ci--4054.1.0--a--b8703b3246-k8s-csi--node--driver--msd5w-eth0" Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.215 [INFO][5327] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:15.219401 containerd[1676]: 2024-09-04 17:43:15.216 [INFO][5321] k8s.go 621: Teardown processing complete. ContainerID="2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5" Sep 4 17:43:15.219401 containerd[1676]: time="2024-09-04T17:43:15.217515842Z" level=info msg="TearDown network for sandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" successfully" Sep 4 17:43:15.230822 containerd[1676]: time="2024-09-04T17:43:15.230783142Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:43:15.230948 containerd[1676]: time="2024-09-04T17:43:15.230855443Z" level=info msg="RemovePodSandbox \"2dd8b9f60d41a20ec0221450cdc5bdba87c6ee8eb4d5c41e5dfc10b80a468fd5\" returns successfully" Sep 4 17:43:24.588470 kubelet[3168]: I0904 17:43:24.588414 3168 topology_manager.go:215] "Topology Admit Handler" podUID="6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57" podNamespace="calico-apiserver" podName="calico-apiserver-7cf79bdfb6-2fks5" Sep 4 17:43:24.596702 kubelet[3168]: W0904 17:43:24.596666 3168 reflector.go:547] object-"calico-apiserver"/"calico-apiserver-certs": failed to list *v1.Secret: secrets "calico-apiserver-certs" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "secrets" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:43:24.596832 kubelet[3168]: E0904 17:43:24.596714 3168 reflector.go:150] object-"calico-apiserver"/"calico-apiserver-certs": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "calico-apiserver-certs" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "secrets" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:43:24.596832 kubelet[3168]: W0904 17:43:24.596821 3168 reflector.go:547] object-"calico-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:43:24.597065 kubelet[3168]: E0904 17:43:24.596837 3168 reflector.go:150] object-"calico-apiserver"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4054.1.0-a-b8703b3246" cannot list resource "configmaps" in API group "" in the namespace "calico-apiserver": no relationship found between node 'ci-4054.1.0-a-b8703b3246' and this object Sep 4 17:43:24.599972 systemd[1]: Created slice kubepods-besteffort-pod6a0e4a0b_9eb8_4d13_9dcc_6a3b1d965d57.slice - libcontainer container kubepods-besteffort-pod6a0e4a0b_9eb8_4d13_9dcc_6a3b1d965d57.slice. Sep 4 17:43:24.605463 kubelet[3168]: I0904 17:43:24.605429 3168 topology_manager.go:215] "Topology Admit Handler" podUID="103b0edb-bec1-464c-9187-e454e0c13a68" podNamespace="calico-apiserver" podName="calico-apiserver-7cf79bdfb6-rgd7f" Sep 4 17:43:24.614037 systemd[1]: Created slice kubepods-besteffort-pod103b0edb_bec1_464c_9187_e454e0c13a68.slice - libcontainer container kubepods-besteffort-pod103b0edb_bec1_464c_9187_e454e0c13a68.slice. Sep 4 17:43:24.686034 kubelet[3168]: I0904 17:43:24.685990 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns7wc\" (UniqueName: \"kubernetes.io/projected/6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57-kube-api-access-ns7wc\") pod \"calico-apiserver-7cf79bdfb6-2fks5\" (UID: \"6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57\") " pod="calico-apiserver/calico-apiserver-7cf79bdfb6-2fks5" Sep 4 17:43:24.686206 kubelet[3168]: I0904 17:43:24.686040 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/103b0edb-bec1-464c-9187-e454e0c13a68-calico-apiserver-certs\") pod \"calico-apiserver-7cf79bdfb6-rgd7f\" (UID: \"103b0edb-bec1-464c-9187-e454e0c13a68\") " pod="calico-apiserver/calico-apiserver-7cf79bdfb6-rgd7f" Sep 4 17:43:24.686206 kubelet[3168]: I0904 17:43:24.686079 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2khd8\" (UniqueName: \"kubernetes.io/projected/103b0edb-bec1-464c-9187-e454e0c13a68-kube-api-access-2khd8\") pod \"calico-apiserver-7cf79bdfb6-rgd7f\" (UID: \"103b0edb-bec1-464c-9187-e454e0c13a68\") " pod="calico-apiserver/calico-apiserver-7cf79bdfb6-rgd7f" Sep 4 17:43:24.686206 kubelet[3168]: I0904 17:43:24.686102 3168 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57-calico-apiserver-certs\") pod \"calico-apiserver-7cf79bdfb6-2fks5\" (UID: \"6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57\") " pod="calico-apiserver/calico-apiserver-7cf79bdfb6-2fks5" Sep 4 17:43:25.805647 containerd[1676]: time="2024-09-04T17:43:25.805596666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7cf79bdfb6-2fks5,Uid:6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57,Namespace:calico-apiserver,Attempt:0,}" Sep 4 17:43:25.820391 containerd[1676]: time="2024-09-04T17:43:25.820349294Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7cf79bdfb6-rgd7f,Uid:103b0edb-bec1-464c-9187-e454e0c13a68,Namespace:calico-apiserver,Attempt:0,}" Sep 4 17:43:25.979014 systemd-networkd[1471]: cali4418b6a7676: Link UP Sep 4 17:43:25.981645 systemd-networkd[1471]: cali4418b6a7676: Gained carrier Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.882 [INFO][5395] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0 calico-apiserver-7cf79bdfb6- calico-apiserver 6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57 838 0 2024-09-04 17:43:24 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7cf79bdfb6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 calico-apiserver-7cf79bdfb6-2fks5 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali4418b6a7676 [] []}} ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.882 [INFO][5395] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.935 [INFO][5417] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" HandleID="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.943 [INFO][5417] ipam_plugin.go 270: Auto assigning IP ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" HandleID="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002917d0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4054.1.0-a-b8703b3246", "pod":"calico-apiserver-7cf79bdfb6-2fks5", "timestamp":"2024-09-04 17:43:25.93507947 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.943 [INFO][5417] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.944 [INFO][5417] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.944 [INFO][5417] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.945 [INFO][5417] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.949 [INFO][5417] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.953 [INFO][5417] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.955 [INFO][5417] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.957 [INFO][5417] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.957 [INFO][5417] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.959 [INFO][5417] ipam.go 1685: Creating new handle: k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6 Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.961 [INFO][5417] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5417] ipam.go 1216: Successfully claimed IPs: [192.168.18.133/26] block=192.168.18.128/26 handle="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5417] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.133/26] handle="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5417] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:26.006767 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5417] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.133/26] IPv6=[] ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" HandleID="k8s-pod-network.a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:25.973 [INFO][5395] k8s.go 386: Populated endpoint ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0", GenerateName:"calico-apiserver-7cf79bdfb6-", Namespace:"calico-apiserver", SelfLink:"", UID:"6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57", ResourceVersion:"838", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7cf79bdfb6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"calico-apiserver-7cf79bdfb6-2fks5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.18.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4418b6a7676", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:25.973 [INFO][5395] k8s.go 387: Calico CNI using IPs: [192.168.18.133/32] ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:25.973 [INFO][5395] dataplane_linux.go 68: Setting the host side veth name to cali4418b6a7676 ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:25.981 [INFO][5395] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:25.983 [INFO][5395] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0", GenerateName:"calico-apiserver-7cf79bdfb6-", Namespace:"calico-apiserver", SelfLink:"", UID:"6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57", ResourceVersion:"838", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7cf79bdfb6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6", Pod:"calico-apiserver-7cf79bdfb6-2fks5", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.18.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4418b6a7676", MAC:"ea:68:d5:61:4d:b4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:26.008244 containerd[1676]: 2024-09-04 17:43:26.001 [INFO][5395] k8s.go 500: Wrote updated endpoint to datastore ContainerID="a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-2fks5" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--2fks5-eth0" Sep 4 17:43:26.043258 systemd-networkd[1471]: cali31cd6b16646: Link UP Sep 4 17:43:26.044428 systemd-networkd[1471]: cali31cd6b16646: Gained carrier Sep 4 17:43:26.058370 containerd[1676]: time="2024-09-04T17:43:26.056216144Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:26.058370 containerd[1676]: time="2024-09-04T17:43:26.056289046Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:26.058370 containerd[1676]: time="2024-09-04T17:43:26.056309046Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:26.058370 containerd[1676]: time="2024-09-04T17:43:26.056624351Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.906 [INFO][5405] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0 calico-apiserver-7cf79bdfb6- calico-apiserver 103b0edb-bec1-464c-9187-e454e0c13a68 843 0 2024-09-04 17:43:24 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7cf79bdfb6 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4054.1.0-a-b8703b3246 calico-apiserver-7cf79bdfb6-rgd7f eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali31cd6b16646 [] []}} ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.907 [INFO][5405] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.948 [INFO][5423] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" HandleID="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.956 [INFO][5423] ipam_plugin.go 270: Auto assigning IP ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" HandleID="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000114a10), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4054.1.0-a-b8703b3246", "pod":"calico-apiserver-7cf79bdfb6-rgd7f", "timestamp":"2024-09-04 17:43:25.948373876 +0000 UTC"}, Hostname:"ci-4054.1.0-a-b8703b3246", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.956 [INFO][5423] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5423] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.968 [INFO][5423] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4054.1.0-a-b8703b3246' Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.970 [INFO][5423] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:25.978 [INFO][5423] ipam.go 372: Looking up existing affinities for host host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.001 [INFO][5423] ipam.go 489: Trying affinity for 192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.003 [INFO][5423] ipam.go 155: Attempting to load block cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.006 [INFO][5423] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.18.128/26 host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.006 [INFO][5423] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.18.128/26 handle="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.008 [INFO][5423] ipam.go 1685: Creating new handle: k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.019 [INFO][5423] ipam.go 1203: Writing block in order to claim IPs block=192.168.18.128/26 handle="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.031 [INFO][5423] ipam.go 1216: Successfully claimed IPs: [192.168.18.134/26] block=192.168.18.128/26 handle="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.031 [INFO][5423] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.18.134/26] handle="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" host="ci-4054.1.0-a-b8703b3246" Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.033 [INFO][5423] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:43:26.068547 containerd[1676]: 2024-09-04 17:43:26.033 [INFO][5423] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.18.134/26] IPv6=[] ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" HandleID="k8s-pod-network.e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Workload="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.038 [INFO][5405] k8s.go 386: Populated endpoint ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0", GenerateName:"calico-apiserver-7cf79bdfb6-", Namespace:"calico-apiserver", SelfLink:"", UID:"103b0edb-bec1-464c-9187-e454e0c13a68", ResourceVersion:"843", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7cf79bdfb6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"", Pod:"calico-apiserver-7cf79bdfb6-rgd7f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.18.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali31cd6b16646", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.039 [INFO][5405] k8s.go 387: Calico CNI using IPs: [192.168.18.134/32] ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.039 [INFO][5405] dataplane_linux.go 68: Setting the host side veth name to cali31cd6b16646 ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.043 [INFO][5405] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.044 [INFO][5405] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0", GenerateName:"calico-apiserver-7cf79bdfb6-", Namespace:"calico-apiserver", SelfLink:"", UID:"103b0edb-bec1-464c-9187-e454e0c13a68", ResourceVersion:"843", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7cf79bdfb6", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4054.1.0-a-b8703b3246", ContainerID:"e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca", Pod:"calico-apiserver-7cf79bdfb6-rgd7f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.18.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali31cd6b16646", MAC:"ae:bc:0b:7c:f7:33", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:43:26.070619 containerd[1676]: 2024-09-04 17:43:26.056 [INFO][5405] k8s.go 500: Wrote updated endpoint to datastore ContainerID="e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca" Namespace="calico-apiserver" Pod="calico-apiserver-7cf79bdfb6-rgd7f" WorkloadEndpoint="ci--4054.1.0--a--b8703b3246-k8s-calico--apiserver--7cf79bdfb6--rgd7f-eth0" Sep 4 17:43:26.107324 containerd[1676]: time="2024-09-04T17:43:26.106801427Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:43:26.107324 containerd[1676]: time="2024-09-04T17:43:26.107276835Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:43:26.107962 containerd[1676]: time="2024-09-04T17:43:26.107765742Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:26.108088 systemd[1]: Started cri-containerd-a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6.scope - libcontainer container a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6. Sep 4 17:43:26.108777 containerd[1676]: time="2024-09-04T17:43:26.107906644Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:43:26.132082 systemd[1]: Started cri-containerd-e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca.scope - libcontainer container e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca. Sep 4 17:43:26.192478 containerd[1676]: time="2024-09-04T17:43:26.192238149Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7cf79bdfb6-rgd7f,Uid:103b0edb-bec1-464c-9187-e454e0c13a68,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca\"" Sep 4 17:43:26.197380 containerd[1676]: time="2024-09-04T17:43:26.197070724Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\"" Sep 4 17:43:26.209844 containerd[1676]: time="2024-09-04T17:43:26.209747520Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7cf79bdfb6-2fks5,Uid:6a0e4a0b-9eb8-4d13-9dcc-6a3b1d965d57,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6\"" Sep 4 17:43:27.282055 systemd-networkd[1471]: cali31cd6b16646: Gained IPv6LL Sep 4 17:43:27.602037 systemd-networkd[1471]: cali4418b6a7676: Gained IPv6LL Sep 4 17:43:28.496899 containerd[1676]: time="2024-09-04T17:43:28.496828014Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:28.499197 containerd[1676]: time="2024-09-04T17:43:28.499130750Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.28.1: active requests=0, bytes read=40419849" Sep 4 17:43:28.502524 containerd[1676]: time="2024-09-04T17:43:28.502467001Z" level=info msg="ImageCreate event name:\"sha256:91dd0fd3dab3f170b52404ec5e67926439207bf71c08b7f54de8f3db6209537b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:28.507801 containerd[1676]: time="2024-09-04T17:43:28.507723282Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b4ee1aa27bdeddc34dd200145eb033b716cf598570206c96693a35a317ab4f1e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:28.508954 containerd[1676]: time="2024-09-04T17:43:28.508379693Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" with image id \"sha256:91dd0fd3dab3f170b52404ec5e67926439207bf71c08b7f54de8f3db6209537b\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b4ee1aa27bdeddc34dd200145eb033b716cf598570206c96693a35a317ab4f1e\", size \"41912266\" in 2.309804145s" Sep 4 17:43:28.508954 containerd[1676]: time="2024-09-04T17:43:28.508421193Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" returns image reference \"sha256:91dd0fd3dab3f170b52404ec5e67926439207bf71c08b7f54de8f3db6209537b\"" Sep 4 17:43:28.509764 containerd[1676]: time="2024-09-04T17:43:28.509443709Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\"" Sep 4 17:43:28.511320 containerd[1676]: time="2024-09-04T17:43:28.511206736Z" level=info msg="CreateContainer within sandbox \"e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 4 17:43:28.549058 containerd[1676]: time="2024-09-04T17:43:28.549010621Z" level=info msg="CreateContainer within sandbox \"e1717dd31dd0f72859019a1d2d20e870c552d660eaaa0944d8058652ebb4ebca\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae\"" Sep 4 17:43:28.551064 containerd[1676]: time="2024-09-04T17:43:28.549586630Z" level=info msg="StartContainer for \"03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae\"" Sep 4 17:43:28.585555 systemd[1]: run-containerd-runc-k8s.io-03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae-runc.shEFXM.mount: Deactivated successfully. Sep 4 17:43:28.598056 systemd[1]: Started cri-containerd-03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae.scope - libcontainer container 03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae. Sep 4 17:43:28.649531 containerd[1676]: time="2024-09-04T17:43:28.649480876Z" level=info msg="StartContainer for \"03897e108eebc1fd36326dbe6ef433779733b0e78306a681a1ed60ba5b9cccae\" returns successfully" Sep 4 17:43:28.839010 containerd[1676]: time="2024-09-04T17:43:28.838229197Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:43:28.842053 containerd[1676]: time="2024-09-04T17:43:28.842008556Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.28.1: active requests=0, bytes read=77" Sep 4 17:43:28.844632 containerd[1676]: time="2024-09-04T17:43:28.844591096Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" with image id \"sha256:91dd0fd3dab3f170b52404ec5e67926439207bf71c08b7f54de8f3db6209537b\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b4ee1aa27bdeddc34dd200145eb033b716cf598570206c96693a35a317ab4f1e\", size \"41912266\" in 335.114186ms" Sep 4 17:43:28.844751 containerd[1676]: time="2024-09-04T17:43:28.844734398Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" returns image reference \"sha256:91dd0fd3dab3f170b52404ec5e67926439207bf71c08b7f54de8f3db6209537b\"" Sep 4 17:43:28.849366 containerd[1676]: time="2024-09-04T17:43:28.848794061Z" level=info msg="CreateContainer within sandbox \"a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 4 17:43:28.881273 containerd[1676]: time="2024-09-04T17:43:28.881230863Z" level=info msg="CreateContainer within sandbox \"a7455773864b4bab0ff784b56d0c39b7d11c9b3311c6d609e5bad46235496bd6\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"cee8fbd82883fb2cb2c9d8ae0e99a440de1d1f2747515458746ddbe4b8aa910a\"" Sep 4 17:43:28.883570 containerd[1676]: time="2024-09-04T17:43:28.883526998Z" level=info msg="StartContainer for \"cee8fbd82883fb2cb2c9d8ae0e99a440de1d1f2747515458746ddbe4b8aa910a\"" Sep 4 17:43:28.931111 systemd[1]: Started cri-containerd-cee8fbd82883fb2cb2c9d8ae0e99a440de1d1f2747515458746ddbe4b8aa910a.scope - libcontainer container cee8fbd82883fb2cb2c9d8ae0e99a440de1d1f2747515458746ddbe4b8aa910a. Sep 4 17:43:29.020906 kubelet[3168]: I0904 17:43:29.020833 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7cf79bdfb6-rgd7f" podStartSLOduration=2.706384106 podStartE2EDuration="5.020813423s" podCreationTimestamp="2024-09-04 17:43:24 +0000 UTC" firstStartedPulling="2024-09-04 17:43:26.19484629 +0000 UTC m=+71.568197632" lastFinishedPulling="2024-09-04 17:43:28.509275507 +0000 UTC m=+73.882626949" observedRunningTime="2024-09-04 17:43:29.019589804 +0000 UTC m=+74.392941246" watchObservedRunningTime="2024-09-04 17:43:29.020813423 +0000 UTC m=+74.394164765" Sep 4 17:43:30.434181 containerd[1676]: time="2024-09-04T17:43:30.434111994Z" level=info msg="StartContainer for \"cee8fbd82883fb2cb2c9d8ae0e99a440de1d1f2747515458746ddbe4b8aa910a\" returns successfully" Sep 4 17:43:31.453998 kubelet[3168]: I0904 17:43:31.453426 3168 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7cf79bdfb6-2fks5" podStartSLOduration=4.818401984 podStartE2EDuration="7.453403562s" podCreationTimestamp="2024-09-04 17:43:24 +0000 UTC" firstStartedPulling="2024-09-04 17:43:26.211095641 +0000 UTC m=+71.584447083" lastFinishedPulling="2024-09-04 17:43:28.846097219 +0000 UTC m=+74.219448661" observedRunningTime="2024-09-04 17:43:30.450164643 +0000 UTC m=+75.823515985" watchObservedRunningTime="2024-09-04 17:43:31.453403562 +0000 UTC m=+76.826755004" Sep 4 17:43:53.375662 systemd[1]: run-containerd-runc-k8s.io-e549d05cb9b8618cc91d10b608dff461a119c44d6d13a186be377db14ea7386d-runc.JikHzr.mount: Deactivated successfully. Sep 4 17:44:14.568418 systemd[1]: Started sshd@7-10.200.4.34:22-10.200.16.10:57694.service - OpenSSH per-connection server daemon (10.200.16.10:57694). Sep 4 17:44:15.172481 sshd[5725]: Accepted publickey for core from 10.200.16.10 port 57694 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:15.174363 sshd[5725]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:15.179214 systemd-logind[1648]: New session 10 of user core. Sep 4 17:44:15.183039 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 4 17:44:15.654359 sshd[5725]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:15.657966 systemd[1]: sshd@7-10.200.4.34:22-10.200.16.10:57694.service: Deactivated successfully. Sep 4 17:44:15.660627 systemd[1]: session-10.scope: Deactivated successfully. Sep 4 17:44:15.662669 systemd-logind[1648]: Session 10 logged out. Waiting for processes to exit. Sep 4 17:44:15.664222 systemd-logind[1648]: Removed session 10. Sep 4 17:44:20.758240 systemd[1]: Started sshd@8-10.200.4.34:22-10.200.16.10:45504.service - OpenSSH per-connection server daemon (10.200.16.10:45504). Sep 4 17:44:21.344109 sshd[5783]: Accepted publickey for core from 10.200.16.10 port 45504 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:21.345611 sshd[5783]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:21.350288 systemd-logind[1648]: New session 11 of user core. Sep 4 17:44:21.357045 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 4 17:44:21.816994 sshd[5783]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:21.820461 systemd[1]: sshd@8-10.200.4.34:22-10.200.16.10:45504.service: Deactivated successfully. Sep 4 17:44:21.823035 systemd[1]: session-11.scope: Deactivated successfully. Sep 4 17:44:21.825069 systemd-logind[1648]: Session 11 logged out. Waiting for processes to exit. Sep 4 17:44:21.826098 systemd-logind[1648]: Removed session 11. Sep 4 17:44:26.924192 systemd[1]: Started sshd@9-10.200.4.34:22-10.200.16.10:45510.service - OpenSSH per-connection server daemon (10.200.16.10:45510). Sep 4 17:44:27.505848 sshd[5818]: Accepted publickey for core from 10.200.16.10 port 45510 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:27.507685 sshd[5818]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:27.513548 systemd-logind[1648]: New session 12 of user core. Sep 4 17:44:27.518056 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 4 17:44:27.991619 sshd[5818]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:27.994690 systemd[1]: sshd@9-10.200.4.34:22-10.200.16.10:45510.service: Deactivated successfully. Sep 4 17:44:27.997137 systemd[1]: session-12.scope: Deactivated successfully. Sep 4 17:44:27.998696 systemd-logind[1648]: Session 12 logged out. Waiting for processes to exit. Sep 4 17:44:27.999997 systemd-logind[1648]: Removed session 12. Sep 4 17:44:28.101218 systemd[1]: Started sshd@10-10.200.4.34:22-10.200.16.10:45522.service - OpenSSH per-connection server daemon (10.200.16.10:45522). Sep 4 17:44:28.708030 sshd[5838]: Accepted publickey for core from 10.200.16.10 port 45522 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:28.710350 sshd[5838]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:28.717102 systemd-logind[1648]: New session 13 of user core. Sep 4 17:44:28.722075 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 4 17:44:29.222796 sshd[5838]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:29.226012 systemd[1]: sshd@10-10.200.4.34:22-10.200.16.10:45522.service: Deactivated successfully. Sep 4 17:44:29.228261 systemd[1]: session-13.scope: Deactivated successfully. Sep 4 17:44:29.229824 systemd-logind[1648]: Session 13 logged out. Waiting for processes to exit. Sep 4 17:44:29.231369 systemd-logind[1648]: Removed session 13. Sep 4 17:44:29.332185 systemd[1]: Started sshd@11-10.200.4.34:22-10.200.16.10:43270.service - OpenSSH per-connection server daemon (10.200.16.10:43270). Sep 4 17:44:29.923820 sshd[5849]: Accepted publickey for core from 10.200.16.10 port 43270 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:29.925776 sshd[5849]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:29.929984 systemd-logind[1648]: New session 14 of user core. Sep 4 17:44:29.935061 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 4 17:44:30.412783 sshd[5849]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:30.416346 systemd[1]: sshd@11-10.200.4.34:22-10.200.16.10:43270.service: Deactivated successfully. Sep 4 17:44:30.418723 systemd[1]: session-14.scope: Deactivated successfully. Sep 4 17:44:30.420390 systemd-logind[1648]: Session 14 logged out. Waiting for processes to exit. Sep 4 17:44:30.421594 systemd-logind[1648]: Removed session 14. Sep 4 17:44:35.523858 systemd[1]: Started sshd@12-10.200.4.34:22-10.200.16.10:43274.service - OpenSSH per-connection server daemon (10.200.16.10:43274). Sep 4 17:44:36.108977 sshd[5869]: Accepted publickey for core from 10.200.16.10 port 43274 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:36.110633 sshd[5869]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:36.115388 systemd-logind[1648]: New session 15 of user core. Sep 4 17:44:36.121097 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 4 17:44:36.604109 sshd[5869]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:36.609226 systemd-logind[1648]: Session 15 logged out. Waiting for processes to exit. Sep 4 17:44:36.609642 systemd[1]: sshd@12-10.200.4.34:22-10.200.16.10:43274.service: Deactivated successfully. Sep 4 17:44:36.612409 systemd[1]: session-15.scope: Deactivated successfully. Sep 4 17:44:36.614328 systemd-logind[1648]: Removed session 15. Sep 4 17:44:41.711863 systemd[1]: Started sshd@13-10.200.4.34:22-10.200.16.10:40206.service - OpenSSH per-connection server daemon (10.200.16.10:40206). Sep 4 17:44:42.297489 sshd[5890]: Accepted publickey for core from 10.200.16.10 port 40206 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:42.299396 sshd[5890]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:42.303981 systemd-logind[1648]: New session 16 of user core. Sep 4 17:44:42.311036 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 4 17:44:42.771105 sshd[5890]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:42.775714 systemd[1]: sshd@13-10.200.4.34:22-10.200.16.10:40206.service: Deactivated successfully. Sep 4 17:44:42.778362 systemd[1]: session-16.scope: Deactivated successfully. Sep 4 17:44:42.779564 systemd-logind[1648]: Session 16 logged out. Waiting for processes to exit. Sep 4 17:44:42.780779 systemd-logind[1648]: Removed session 16. Sep 4 17:44:47.866925 systemd[1]: run-containerd-runc-k8s.io-2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e-runc.bcJwWY.mount: Deactivated successfully. Sep 4 17:44:47.887986 systemd[1]: Started sshd@14-10.200.4.34:22-10.200.16.10:40210.service - OpenSSH per-connection server daemon (10.200.16.10:40210). Sep 4 17:44:48.472923 sshd[5933]: Accepted publickey for core from 10.200.16.10 port 40210 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:48.474741 sshd[5933]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:48.479744 systemd-logind[1648]: New session 17 of user core. Sep 4 17:44:48.486058 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 4 17:44:48.957067 sshd[5933]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:48.960634 systemd[1]: sshd@14-10.200.4.34:22-10.200.16.10:40210.service: Deactivated successfully. Sep 4 17:44:48.963342 systemd[1]: session-17.scope: Deactivated successfully. Sep 4 17:44:48.965366 systemd-logind[1648]: Session 17 logged out. Waiting for processes to exit. Sep 4 17:44:48.966622 systemd-logind[1648]: Removed session 17. Sep 4 17:44:54.068178 systemd[1]: Started sshd@15-10.200.4.34:22-10.200.16.10:42050.service - OpenSSH per-connection server daemon (10.200.16.10:42050). Sep 4 17:44:54.656316 sshd[5976]: Accepted publickey for core from 10.200.16.10 port 42050 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:54.657839 sshd[5976]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:54.661940 systemd-logind[1648]: New session 18 of user core. Sep 4 17:44:54.667293 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 4 17:44:55.139465 sshd[5976]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:55.143185 systemd[1]: sshd@15-10.200.4.34:22-10.200.16.10:42050.service: Deactivated successfully. Sep 4 17:44:55.145828 systemd[1]: session-18.scope: Deactivated successfully. Sep 4 17:44:55.148122 systemd-logind[1648]: Session 18 logged out. Waiting for processes to exit. Sep 4 17:44:55.149578 systemd-logind[1648]: Removed session 18. Sep 4 17:44:55.248260 systemd[1]: Started sshd@16-10.200.4.34:22-10.200.16.10:42062.service - OpenSSH per-connection server daemon (10.200.16.10:42062). Sep 4 17:44:55.866606 sshd[5989]: Accepted publickey for core from 10.200.16.10 port 42062 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:55.868430 sshd[5989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:55.874358 systemd-logind[1648]: New session 19 of user core. Sep 4 17:44:55.880061 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 4 17:44:56.380418 sshd[5989]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:56.385222 systemd[1]: sshd@16-10.200.4.34:22-10.200.16.10:42062.service: Deactivated successfully. Sep 4 17:44:56.387440 systemd[1]: session-19.scope: Deactivated successfully. Sep 4 17:44:56.388310 systemd-logind[1648]: Session 19 logged out. Waiting for processes to exit. Sep 4 17:44:56.389403 systemd-logind[1648]: Removed session 19. Sep 4 17:44:56.484208 systemd[1]: Started sshd@17-10.200.4.34:22-10.200.16.10:42078.service - OpenSSH per-connection server daemon (10.200.16.10:42078). Sep 4 17:44:57.067115 sshd[5999]: Accepted publickey for core from 10.200.16.10 port 42078 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:44:57.068927 sshd[5999]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:44:57.073754 systemd-logind[1648]: New session 20 of user core. Sep 4 17:44:57.080038 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 4 17:44:59.338256 sshd[5999]: pam_unix(sshd:session): session closed for user core Sep 4 17:44:59.342100 systemd-logind[1648]: Session 20 logged out. Waiting for processes to exit. Sep 4 17:44:59.345040 systemd[1]: sshd@17-10.200.4.34:22-10.200.16.10:42078.service: Deactivated successfully. Sep 4 17:44:59.347669 systemd[1]: session-20.scope: Deactivated successfully. Sep 4 17:44:59.349195 systemd-logind[1648]: Removed session 20. Sep 4 17:44:59.450191 systemd[1]: Started sshd@18-10.200.4.34:22-10.200.16.10:49110.service - OpenSSH per-connection server daemon (10.200.16.10:49110). Sep 4 17:45:00.036852 sshd[6017]: Accepted publickey for core from 10.200.16.10 port 49110 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:00.038876 sshd[6017]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:00.043761 systemd-logind[1648]: New session 21 of user core. Sep 4 17:45:00.049068 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 4 17:45:00.612587 sshd[6017]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:00.615601 systemd[1]: sshd@18-10.200.4.34:22-10.200.16.10:49110.service: Deactivated successfully. Sep 4 17:45:00.617999 systemd[1]: session-21.scope: Deactivated successfully. Sep 4 17:45:00.619812 systemd-logind[1648]: Session 21 logged out. Waiting for processes to exit. Sep 4 17:45:00.621043 systemd-logind[1648]: Removed session 21. Sep 4 17:45:00.731014 systemd[1]: Started sshd@19-10.200.4.34:22-10.200.16.10:49112.service - OpenSSH per-connection server daemon (10.200.16.10:49112). Sep 4 17:45:01.315762 sshd[6030]: Accepted publickey for core from 10.200.16.10 port 49112 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:01.317363 sshd[6030]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:01.321492 systemd-logind[1648]: New session 22 of user core. Sep 4 17:45:01.331053 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 4 17:45:01.806264 sshd[6030]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:01.810333 systemd[1]: sshd@19-10.200.4.34:22-10.200.16.10:49112.service: Deactivated successfully. Sep 4 17:45:01.812473 systemd[1]: session-22.scope: Deactivated successfully. Sep 4 17:45:01.813468 systemd-logind[1648]: Session 22 logged out. Waiting for processes to exit. Sep 4 17:45:01.814575 systemd-logind[1648]: Removed session 22. Sep 4 17:45:06.913189 systemd[1]: Started sshd@20-10.200.4.34:22-10.200.16.10:49128.service - OpenSSH per-connection server daemon (10.200.16.10:49128). Sep 4 17:45:07.498439 sshd[6048]: Accepted publickey for core from 10.200.16.10 port 49128 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:07.500308 sshd[6048]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:07.505660 systemd-logind[1648]: New session 23 of user core. Sep 4 17:45:07.511061 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 4 17:45:07.978674 sshd[6048]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:07.983433 systemd[1]: sshd@20-10.200.4.34:22-10.200.16.10:49128.service: Deactivated successfully. Sep 4 17:45:07.985815 systemd[1]: session-23.scope: Deactivated successfully. Sep 4 17:45:07.986635 systemd-logind[1648]: Session 23 logged out. Waiting for processes to exit. Sep 4 17:45:07.987719 systemd-logind[1648]: Removed session 23. Sep 4 17:45:13.088230 systemd[1]: Started sshd@21-10.200.4.34:22-10.200.16.10:57568.service - OpenSSH per-connection server daemon (10.200.16.10:57568). Sep 4 17:45:13.669601 sshd[6068]: Accepted publickey for core from 10.200.16.10 port 57568 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:13.672432 sshd[6068]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:13.677379 systemd-logind[1648]: New session 24 of user core. Sep 4 17:45:13.683062 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 4 17:45:14.151631 sshd[6068]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:14.155285 systemd[1]: sshd@21-10.200.4.34:22-10.200.16.10:57568.service: Deactivated successfully. Sep 4 17:45:14.157619 systemd[1]: session-24.scope: Deactivated successfully. Sep 4 17:45:14.159434 systemd-logind[1648]: Session 24 logged out. Waiting for processes to exit. Sep 4 17:45:14.160570 systemd-logind[1648]: Removed session 24. Sep 4 17:45:17.865098 systemd[1]: run-containerd-runc-k8s.io-2dde28e9072a5e64067e121080d75607ba170ee5995d24bfa2bb6222c07be03e-runc.h7Fcou.mount: Deactivated successfully. Sep 4 17:45:19.254852 systemd[1]: Started sshd@22-10.200.4.34:22-10.200.16.10:42426.service - OpenSSH per-connection server daemon (10.200.16.10:42426). Sep 4 17:45:19.846470 sshd[6120]: Accepted publickey for core from 10.200.16.10 port 42426 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:19.848527 sshd[6120]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:19.852994 systemd-logind[1648]: New session 25 of user core. Sep 4 17:45:19.856086 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 4 17:45:20.313786 sshd[6120]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:20.317570 systemd[1]: sshd@22-10.200.4.34:22-10.200.16.10:42426.service: Deactivated successfully. Sep 4 17:45:20.320104 systemd[1]: session-25.scope: Deactivated successfully. Sep 4 17:45:20.321632 systemd-logind[1648]: Session 25 logged out. Waiting for processes to exit. Sep 4 17:45:20.323026 systemd-logind[1648]: Removed session 25. Sep 4 17:45:25.423149 systemd[1]: Started sshd@23-10.200.4.34:22-10.200.16.10:42438.service - OpenSSH per-connection server daemon (10.200.16.10:42438). Sep 4 17:45:26.023713 sshd[6160]: Accepted publickey for core from 10.200.16.10 port 42438 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:26.025236 sshd[6160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:26.029998 systemd-logind[1648]: New session 26 of user core. Sep 4 17:45:26.038045 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 4 17:45:26.535660 sshd[6160]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:26.539835 systemd[1]: sshd@23-10.200.4.34:22-10.200.16.10:42438.service: Deactivated successfully. Sep 4 17:45:26.542140 systemd[1]: session-26.scope: Deactivated successfully. Sep 4 17:45:26.542987 systemd-logind[1648]: Session 26 logged out. Waiting for processes to exit. Sep 4 17:45:26.543938 systemd-logind[1648]: Removed session 26. Sep 4 17:45:31.648184 systemd[1]: Started sshd@24-10.200.4.34:22-10.200.16.10:39596.service - OpenSSH per-connection server daemon (10.200.16.10:39596). Sep 4 17:45:32.225090 sshd[6175]: Accepted publickey for core from 10.200.16.10 port 39596 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:32.226579 sshd[6175]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:32.231346 systemd-logind[1648]: New session 27 of user core. Sep 4 17:45:32.238055 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 4 17:45:32.694248 sshd[6175]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:32.698286 systemd[1]: sshd@24-10.200.4.34:22-10.200.16.10:39596.service: Deactivated successfully. Sep 4 17:45:32.700430 systemd[1]: session-27.scope: Deactivated successfully. Sep 4 17:45:32.701237 systemd-logind[1648]: Session 27 logged out. Waiting for processes to exit. Sep 4 17:45:32.702420 systemd-logind[1648]: Removed session 27. Sep 4 17:45:37.805210 systemd[1]: Started sshd@25-10.200.4.34:22-10.200.16.10:39598.service - OpenSSH per-connection server daemon (10.200.16.10:39598). Sep 4 17:45:38.391018 sshd[6192]: Accepted publickey for core from 10.200.16.10 port 39598 ssh2: RSA SHA256:Uyt7oO2EXSFubEx3In16nTY26l+8pigGedEy8tt2QcY Sep 4 17:45:38.392533 sshd[6192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:45:38.397169 systemd-logind[1648]: New session 28 of user core. Sep 4 17:45:38.404047 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 4 17:45:38.876485 sshd[6192]: pam_unix(sshd:session): session closed for user core Sep 4 17:45:38.881244 systemd[1]: sshd@25-10.200.4.34:22-10.200.16.10:39598.service: Deactivated successfully. Sep 4 17:45:38.883784 systemd[1]: session-28.scope: Deactivated successfully. Sep 4 17:45:38.885051 systemd-logind[1648]: Session 28 logged out. Waiting for processes to exit. Sep 4 17:45:38.886309 systemd-logind[1648]: Removed session 28.