Feb 13 15:36:55.010744 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 13:54:58 -00 2025 Feb 13 15:36:55.010786 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=cd73eba291b8356dfc2c39f651cabef9206685f772c8949188fd366788d672c2 Feb 13 15:36:55.010806 kernel: BIOS-provided physical RAM map: Feb 13 15:36:55.010816 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Feb 13 15:36:55.010824 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Feb 13 15:36:55.010835 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Feb 13 15:36:55.010850 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007d9e9fff] usable Feb 13 15:36:55.010861 kernel: BIOS-e820: [mem 0x000000007d9ea000-0x000000007fffffff] reserved Feb 13 15:36:55.010873 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000e03fffff] reserved Feb 13 15:36:55.011006 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Feb 13 15:36:55.011019 kernel: NX (Execute Disable) protection: active Feb 13 15:36:55.011031 kernel: APIC: Static calls initialized Feb 13 15:36:55.011042 kernel: SMBIOS 2.7 present. Feb 13 15:36:55.011054 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Feb 13 15:36:55.011075 kernel: Hypervisor detected: KVM Feb 13 15:36:55.011086 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 13 15:36:55.011099 kernel: kvm-clock: using sched offset of 7912743692 cycles Feb 13 15:36:55.011113 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 13 15:36:55.011142 kernel: tsc: Detected 2499.998 MHz processor Feb 13 15:36:55.011157 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 15:36:55.011172 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 15:36:55.011190 kernel: last_pfn = 0x7d9ea max_arch_pfn = 0x400000000 Feb 13 15:36:55.011205 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Feb 13 15:36:55.011219 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 15:36:55.011233 kernel: Using GB pages for direct mapping Feb 13 15:36:55.011247 kernel: ACPI: Early table checksum verification disabled Feb 13 15:36:55.011261 kernel: ACPI: RSDP 0x00000000000F8F40 000014 (v00 AMAZON) Feb 13 15:36:55.011276 kernel: ACPI: RSDT 0x000000007D9EE350 000044 (v01 AMAZON AMZNRSDT 00000001 AMZN 00000001) Feb 13 15:36:55.011290 kernel: ACPI: FACP 0x000000007D9EFF80 000074 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Feb 13 15:36:55.011304 kernel: ACPI: DSDT 0x000000007D9EE3A0 0010E9 (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Feb 13 15:36:55.011321 kernel: ACPI: FACS 0x000000007D9EFF40 000040 Feb 13 15:36:55.011336 kernel: ACPI: SSDT 0x000000007D9EF6C0 00087A (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Feb 13 15:36:55.011350 kernel: ACPI: APIC 0x000000007D9EF5D0 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Feb 13 15:36:55.011364 kernel: ACPI: SRAT 0x000000007D9EF530 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Feb 13 15:36:55.011379 kernel: ACPI: SLIT 0x000000007D9EF4C0 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Feb 13 15:36:55.011393 kernel: ACPI: WAET 0x000000007D9EF490 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Feb 13 15:36:55.011407 kernel: ACPI: HPET 0x00000000000C9000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Feb 13 15:36:55.011421 kernel: ACPI: SSDT 0x00000000000C9040 00007B (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Feb 13 15:36:55.011436 kernel: ACPI: Reserving FACP table memory at [mem 0x7d9eff80-0x7d9efff3] Feb 13 15:36:55.011454 kernel: ACPI: Reserving DSDT table memory at [mem 0x7d9ee3a0-0x7d9ef488] Feb 13 15:36:55.011473 kernel: ACPI: Reserving FACS table memory at [mem 0x7d9eff40-0x7d9eff7f] Feb 13 15:36:55.011586 kernel: ACPI: Reserving SSDT table memory at [mem 0x7d9ef6c0-0x7d9eff39] Feb 13 15:36:55.011601 kernel: ACPI: Reserving APIC table memory at [mem 0x7d9ef5d0-0x7d9ef645] Feb 13 15:36:55.011615 kernel: ACPI: Reserving SRAT table memory at [mem 0x7d9ef530-0x7d9ef5cf] Feb 13 15:36:55.011632 kernel: ACPI: Reserving SLIT table memory at [mem 0x7d9ef4c0-0x7d9ef52b] Feb 13 15:36:55.011646 kernel: ACPI: Reserving WAET table memory at [mem 0x7d9ef490-0x7d9ef4b7] Feb 13 15:36:55.011659 kernel: ACPI: Reserving HPET table memory at [mem 0xc9000-0xc9037] Feb 13 15:36:55.011673 kernel: ACPI: Reserving SSDT table memory at [mem 0xc9040-0xc90ba] Feb 13 15:36:55.011688 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 13 15:36:55.011703 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 13 15:36:55.011717 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Feb 13 15:36:55.011732 kernel: NUMA: Initialized distance table, cnt=1 Feb 13 15:36:55.011746 kernel: NODE_DATA(0) allocated [mem 0x7d9e3000-0x7d9e8fff] Feb 13 15:36:55.011764 kernel: Zone ranges: Feb 13 15:36:55.011779 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 15:36:55.011793 kernel: DMA32 [mem 0x0000000001000000-0x000000007d9e9fff] Feb 13 15:36:55.011808 kernel: Normal empty Feb 13 15:36:55.011822 kernel: Movable zone start for each node Feb 13 15:36:55.011836 kernel: Early memory node ranges Feb 13 15:36:55.011849 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Feb 13 15:36:55.011863 kernel: node 0: [mem 0x0000000000100000-0x000000007d9e9fff] Feb 13 15:36:55.011878 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007d9e9fff] Feb 13 15:36:55.011892 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 15:36:55.011910 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Feb 13 15:36:55.011924 kernel: On node 0, zone DMA32: 9750 pages in unavailable ranges Feb 13 15:36:55.011938 kernel: ACPI: PM-Timer IO Port: 0xb008 Feb 13 15:36:55.011953 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 13 15:36:55.011967 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Feb 13 15:36:55.011982 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 13 15:36:55.012122 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 15:36:55.012162 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 13 15:36:55.012177 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 13 15:36:55.012195 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 15:36:55.012210 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 13 15:36:55.012224 kernel: TSC deadline timer available Feb 13 15:36:55.012239 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 13 15:36:55.012253 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Feb 13 15:36:55.012267 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Feb 13 15:36:55.012281 kernel: Booting paravirtualized kernel on KVM Feb 13 15:36:55.012296 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 15:36:55.012311 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Feb 13 15:36:55.012329 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Feb 13 15:36:55.012343 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Feb 13 15:36:55.012357 kernel: pcpu-alloc: [0] 0 1 Feb 13 15:36:55.012371 kernel: kvm-guest: PV spinlocks enabled Feb 13 15:36:55.012385 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 15:36:55.012401 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=cd73eba291b8356dfc2c39f651cabef9206685f772c8949188fd366788d672c2 Feb 13 15:36:55.012416 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 15:36:55.012430 kernel: random: crng init done Feb 13 15:36:55.012447 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 15:36:55.012462 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 13 15:36:55.012476 kernel: Fallback order for Node 0: 0 Feb 13 15:36:55.012490 kernel: Built 1 zonelists, mobility grouping on. Total pages: 506242 Feb 13 15:36:55.012505 kernel: Policy zone: DMA32 Feb 13 15:36:55.012519 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 15:36:55.012533 kernel: Memory: 1932348K/2057760K available (12288K kernel code, 2299K rwdata, 22736K rodata, 42976K init, 2216K bss, 125152K reserved, 0K cma-reserved) Feb 13 15:36:55.012548 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 15:36:55.012562 kernel: Kernel/User page tables isolation: enabled Feb 13 15:36:55.012580 kernel: ftrace: allocating 37920 entries in 149 pages Feb 13 15:36:55.012594 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 15:36:55.012608 kernel: Dynamic Preempt: voluntary Feb 13 15:36:55.012623 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 15:36:55.012639 kernel: rcu: RCU event tracing is enabled. Feb 13 15:36:55.012653 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 15:36:55.012668 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 15:36:55.012682 kernel: Rude variant of Tasks RCU enabled. Feb 13 15:36:55.012696 kernel: Tracing variant of Tasks RCU enabled. Feb 13 15:36:55.012713 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 15:36:55.012727 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 15:36:55.012741 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Feb 13 15:36:55.012754 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 15:36:55.012766 kernel: Console: colour VGA+ 80x25 Feb 13 15:36:55.012780 kernel: printk: console [ttyS0] enabled Feb 13 15:36:55.012794 kernel: ACPI: Core revision 20230628 Feb 13 15:36:55.012809 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Feb 13 15:36:55.012822 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 15:36:55.012839 kernel: x2apic enabled Feb 13 15:36:55.012867 kernel: APIC: Switched APIC routing to: physical x2apic Feb 13 15:36:55.012892 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Feb 13 15:36:55.012907 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499998) Feb 13 15:36:55.012922 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Feb 13 15:36:55.012937 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Feb 13 15:36:55.012953 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 15:36:55.012969 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 15:36:55.012982 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 15:36:55.012996 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 15:36:55.013011 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Feb 13 15:36:55.013025 kernel: RETBleed: Vulnerable Feb 13 15:36:55.013045 kernel: Speculative Store Bypass: Vulnerable Feb 13 15:36:55.013061 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 15:36:55.013077 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 15:36:55.013093 kernel: GDS: Unknown: Dependent on hypervisor status Feb 13 15:36:55.013109 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 15:36:55.013145 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 15:36:55.013162 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 15:36:55.013180 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Feb 13 15:36:55.013206 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Feb 13 15:36:55.013222 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Feb 13 15:36:55.013238 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Feb 13 15:36:55.013252 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Feb 13 15:36:55.013268 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Feb 13 15:36:55.013283 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 15:36:55.013298 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Feb 13 15:36:55.013401 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Feb 13 15:36:55.013478 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Feb 13 15:36:55.013495 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Feb 13 15:36:55.013514 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Feb 13 15:36:55.013529 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Feb 13 15:36:55.013545 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Feb 13 15:36:55.013561 kernel: Freeing SMP alternatives memory: 32K Feb 13 15:36:55.013576 kernel: pid_max: default: 32768 minimum: 301 Feb 13 15:36:55.013592 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 15:36:55.013607 kernel: landlock: Up and running. Feb 13 15:36:55.013622 kernel: SELinux: Initializing. Feb 13 15:36:55.013638 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 13 15:36:55.013653 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 13 15:36:55.013669 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Feb 13 15:36:55.013688 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 15:36:55.013704 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 15:36:55.013720 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 15:36:55.013735 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Feb 13 15:36:55.013751 kernel: signal: max sigframe size: 3632 Feb 13 15:36:55.013766 kernel: rcu: Hierarchical SRCU implementation. Feb 13 15:36:55.013783 kernel: rcu: Max phase no-delay instances is 400. Feb 13 15:36:55.013798 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 13 15:36:55.013814 kernel: smp: Bringing up secondary CPUs ... Feb 13 15:36:55.013832 kernel: smpboot: x86: Booting SMP configuration: Feb 13 15:36:55.013848 kernel: .... node #0, CPUs: #1 Feb 13 15:36:55.013864 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Feb 13 15:36:55.013880 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Feb 13 15:36:55.013896 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 15:36:55.013912 kernel: smpboot: Max logical packages: 1 Feb 13 15:36:55.013927 kernel: smpboot: Total of 2 processors activated (9999.99 BogoMIPS) Feb 13 15:36:55.013942 kernel: devtmpfs: initialized Feb 13 15:36:55.013961 kernel: x86/mm: Memory block size: 128MB Feb 13 15:36:55.013977 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 15:36:55.013992 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 15:36:55.014008 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 15:36:55.014024 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 15:36:55.014039 kernel: audit: initializing netlink subsys (disabled) Feb 13 15:36:55.014055 kernel: audit: type=2000 audit(1739461014.258:1): state=initialized audit_enabled=0 res=1 Feb 13 15:36:55.014070 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 15:36:55.014086 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 15:36:55.014104 kernel: cpuidle: using governor menu Feb 13 15:36:55.014119 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 15:36:55.014151 kernel: dca service started, version 1.12.1 Feb 13 15:36:55.014166 kernel: PCI: Using configuration type 1 for base access Feb 13 15:36:55.014181 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 15:36:55.014197 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 15:36:55.014212 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 15:36:55.014228 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 15:36:55.014242 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 15:36:55.014261 kernel: ACPI: Added _OSI(Module Device) Feb 13 15:36:55.014275 kernel: ACPI: Added _OSI(Processor Device) Feb 13 15:36:55.014290 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 15:36:55.014313 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 15:36:55.014328 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Feb 13 15:36:55.014340 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 15:36:55.014356 kernel: ACPI: Interpreter enabled Feb 13 15:36:55.014371 kernel: ACPI: PM: (supports S0 S5) Feb 13 15:36:55.014388 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 15:36:55.014402 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 15:36:55.014421 kernel: PCI: Using E820 reservations for host bridge windows Feb 13 15:36:55.014437 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F Feb 13 15:36:55.014453 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 15:36:55.014686 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Feb 13 15:36:55.014832 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Feb 13 15:36:55.014968 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Feb 13 15:36:55.014988 kernel: acpiphp: Slot [3] registered Feb 13 15:36:55.015009 kernel: acpiphp: Slot [4] registered Feb 13 15:36:55.015026 kernel: acpiphp: Slot [5] registered Feb 13 15:36:55.015043 kernel: acpiphp: Slot [6] registered Feb 13 15:36:55.015059 kernel: acpiphp: Slot [7] registered Feb 13 15:36:55.015075 kernel: acpiphp: Slot [8] registered Feb 13 15:36:55.015091 kernel: acpiphp: Slot [9] registered Feb 13 15:36:55.015107 kernel: acpiphp: Slot [10] registered Feb 13 15:36:55.015138 kernel: acpiphp: Slot [11] registered Feb 13 15:36:55.015159 kernel: acpiphp: Slot [12] registered Feb 13 15:36:55.015179 kernel: acpiphp: Slot [13] registered Feb 13 15:36:55.015196 kernel: acpiphp: Slot [14] registered Feb 13 15:36:55.015212 kernel: acpiphp: Slot [15] registered Feb 13 15:36:55.015229 kernel: acpiphp: Slot [16] registered Feb 13 15:36:55.015245 kernel: acpiphp: Slot [17] registered Feb 13 15:36:55.015260 kernel: acpiphp: Slot [18] registered Feb 13 15:36:55.015277 kernel: acpiphp: Slot [19] registered Feb 13 15:36:55.015293 kernel: acpiphp: Slot [20] registered Feb 13 15:36:55.015309 kernel: acpiphp: Slot [21] registered Feb 13 15:36:55.015325 kernel: acpiphp: Slot [22] registered Feb 13 15:36:55.015344 kernel: acpiphp: Slot [23] registered Feb 13 15:36:55.015361 kernel: acpiphp: Slot [24] registered Feb 13 15:36:55.015377 kernel: acpiphp: Slot [25] registered Feb 13 15:36:55.015393 kernel: acpiphp: Slot [26] registered Feb 13 15:36:55.015409 kernel: acpiphp: Slot [27] registered Feb 13 15:36:55.015425 kernel: acpiphp: Slot [28] registered Feb 13 15:36:55.015442 kernel: acpiphp: Slot [29] registered Feb 13 15:36:55.015458 kernel: acpiphp: Slot [30] registered Feb 13 15:36:55.015474 kernel: acpiphp: Slot [31] registered Feb 13 15:36:55.015493 kernel: PCI host bridge to bus 0000:00 Feb 13 15:36:55.015639 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 13 15:36:55.015765 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 13 15:36:55.015887 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 13 15:36:55.016007 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Feb 13 15:36:55.016288 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 15:36:55.016450 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Feb 13 15:36:55.016612 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Feb 13 15:36:55.016758 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Feb 13 15:36:55.016898 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Feb 13 15:36:55.017035 kernel: pci 0000:00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 SMB Feb 13 15:36:55.017194 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Feb 13 15:36:55.017334 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Feb 13 15:36:55.017469 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Feb 13 15:36:55.017711 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Feb 13 15:36:55.019435 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Feb 13 15:36:55.019593 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Feb 13 15:36:55.019873 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Feb 13 15:36:55.020019 kernel: pci 0000:00:03.0: reg 0x10: [mem 0xfe400000-0xfe7fffff pref] Feb 13 15:36:55.020270 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Feb 13 15:36:55.020421 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 13 15:36:55.020574 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Feb 13 15:36:55.020831 kernel: pci 0000:00:04.0: reg 0x10: [mem 0xfebf0000-0xfebf3fff] Feb 13 15:36:55.021067 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Feb 13 15:36:55.021264 kernel: pci 0000:00:05.0: reg 0x10: [mem 0xfebf4000-0xfebf7fff] Feb 13 15:36:55.021288 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 13 15:36:55.021305 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 13 15:36:55.021327 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 13 15:36:55.021344 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 13 15:36:55.021361 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Feb 13 15:36:55.021378 kernel: iommu: Default domain type: Translated Feb 13 15:36:55.021394 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 15:36:55.021410 kernel: PCI: Using ACPI for IRQ routing Feb 13 15:36:55.021427 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 13 15:36:55.021443 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Feb 13 15:36:55.021459 kernel: e820: reserve RAM buffer [mem 0x7d9ea000-0x7fffffff] Feb 13 15:36:55.021604 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Feb 13 15:36:55.021740 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Feb 13 15:36:55.021878 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 13 15:36:55.021899 kernel: vgaarb: loaded Feb 13 15:36:55.021917 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Feb 13 15:36:55.021934 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Feb 13 15:36:55.021951 kernel: clocksource: Switched to clocksource kvm-clock Feb 13 15:36:55.021966 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 15:36:55.021983 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 15:36:55.022003 kernel: pnp: PnP ACPI init Feb 13 15:36:55.022020 kernel: pnp: PnP ACPI: found 5 devices Feb 13 15:36:55.022036 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 15:36:55.022052 kernel: NET: Registered PF_INET protocol family Feb 13 15:36:55.022067 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 15:36:55.022083 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Feb 13 15:36:55.022098 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 15:36:55.022116 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 15:36:55.022168 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Feb 13 15:36:55.022185 kernel: TCP: Hash tables configured (established 16384 bind 16384) Feb 13 15:36:55.022201 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 13 15:36:55.022218 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 13 15:36:55.022313 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 15:36:55.022331 kernel: NET: Registered PF_XDP protocol family Feb 13 15:36:55.022483 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 13 15:36:55.022611 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 13 15:36:55.022733 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 13 15:36:55.022860 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Feb 13 15:36:55.023002 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Feb 13 15:36:55.023023 kernel: PCI: CLS 0 bytes, default 64 Feb 13 15:36:55.023041 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 13 15:36:55.023058 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Feb 13 15:36:55.023075 kernel: clocksource: Switched to clocksource tsc Feb 13 15:36:55.023091 kernel: Initialise system trusted keyrings Feb 13 15:36:55.023109 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Feb 13 15:36:55.023161 kernel: Key type asymmetric registered Feb 13 15:36:55.023178 kernel: Asymmetric key parser 'x509' registered Feb 13 15:36:55.023194 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 15:36:55.023210 kernel: io scheduler mq-deadline registered Feb 13 15:36:55.023225 kernel: io scheduler kyber registered Feb 13 15:36:55.023241 kernel: io scheduler bfq registered Feb 13 15:36:55.023257 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 15:36:55.023274 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 15:36:55.023291 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 15:36:55.023311 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 13 15:36:55.023328 kernel: i8042: Warning: Keylock active Feb 13 15:36:55.023345 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 13 15:36:55.023360 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 13 15:36:55.023505 kernel: rtc_cmos 00:00: RTC can wake from S4 Feb 13 15:36:55.023633 kernel: rtc_cmos 00:00: registered as rtc0 Feb 13 15:36:55.023831 kernel: rtc_cmos 00:00: setting system clock to 2025-02-13T15:36:54 UTC (1739461014) Feb 13 15:36:55.023961 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Feb 13 15:36:55.023987 kernel: intel_pstate: CPU model not supported Feb 13 15:36:55.024004 kernel: NET: Registered PF_INET6 protocol family Feb 13 15:36:55.024022 kernel: Segment Routing with IPv6 Feb 13 15:36:55.024038 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 15:36:55.024055 kernel: NET: Registered PF_PACKET protocol family Feb 13 15:36:55.024071 kernel: Key type dns_resolver registered Feb 13 15:36:55.024087 kernel: IPI shorthand broadcast: enabled Feb 13 15:36:55.024104 kernel: sched_clock: Marking stable (571001945, 267330385)->(992312597, -153980267) Feb 13 15:36:55.024121 kernel: registered taskstats version 1 Feb 13 15:36:55.024176 kernel: Loading compiled-in X.509 certificates Feb 13 15:36:55.024193 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: 9ec780e1db69d46be90bbba73ae62b0106e27ae0' Feb 13 15:36:55.024209 kernel: Key type .fscrypt registered Feb 13 15:36:55.024226 kernel: Key type fscrypt-provisioning registered Feb 13 15:36:55.024243 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 15:36:55.024259 kernel: ima: Allocated hash algorithm: sha1 Feb 13 15:36:55.024275 kernel: ima: No architecture policies found Feb 13 15:36:55.024291 kernel: clk: Disabling unused clocks Feb 13 15:36:55.024308 kernel: Freeing unused kernel image (initmem) memory: 42976K Feb 13 15:36:55.024328 kernel: Write protecting the kernel read-only data: 36864k Feb 13 15:36:55.024344 kernel: Freeing unused kernel image (rodata/data gap) memory: 1840K Feb 13 15:36:55.024361 kernel: Run /init as init process Feb 13 15:36:55.024377 kernel: with arguments: Feb 13 15:36:55.024393 kernel: /init Feb 13 15:36:55.024621 kernel: with environment: Feb 13 15:36:55.024644 kernel: HOME=/ Feb 13 15:36:55.024661 kernel: TERM=linux Feb 13 15:36:55.024677 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 15:36:55.024706 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 15:36:55.024741 systemd[1]: Detected virtualization amazon. Feb 13 15:36:55.024763 systemd[1]: Detected architecture x86-64. Feb 13 15:36:55.024780 systemd[1]: Running in initrd. Feb 13 15:36:55.024801 systemd[1]: No hostname configured, using default hostname. Feb 13 15:36:55.024818 systemd[1]: Hostname set to . Feb 13 15:36:55.024837 systemd[1]: Initializing machine ID from VM UUID. Feb 13 15:36:55.024855 systemd[1]: Queued start job for default target initrd.target. Feb 13 15:36:55.024873 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 15:36:55.024892 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 15:36:55.024912 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 15:36:55.024930 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 15:36:55.024951 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 15:36:55.025048 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 15:36:55.025070 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 15:36:55.025088 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 15:36:55.025107 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 15:36:55.025154 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 15:36:55.025173 systemd[1]: Reached target paths.target - Path Units. Feb 13 15:36:55.025222 systemd[1]: Reached target slices.target - Slice Units. Feb 13 15:36:55.025240 systemd[1]: Reached target swap.target - Swaps. Feb 13 15:36:55.025259 systemd[1]: Reached target timers.target - Timer Units. Feb 13 15:36:55.025277 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 15:36:55.025295 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 15:36:55.025314 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 15:36:55.025333 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 15:36:55.025351 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 15:36:55.025369 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 15:36:55.025391 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 15:36:55.025408 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 15:36:55.025427 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 15:36:55.025445 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 15:36:55.025463 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 15:36:55.025481 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 15:36:55.025500 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 15:36:55.025521 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 15:36:55.025574 systemd-journald[179]: Collecting audit messages is disabled. Feb 13 15:36:55.025987 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 13 15:36:55.026015 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 15:36:55.026305 systemd-journald[179]: Journal started Feb 13 15:36:55.026682 systemd-journald[179]: Runtime Journal (/run/log/journal/ec2668ca7830d0977d2a9a726ac7b55a) is 4.8M, max 38.6M, 33.7M free. Feb 13 15:36:55.010841 systemd-modules-load[180]: Inserted module 'overlay' Feb 13 15:36:55.034307 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 15:36:55.040655 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 15:36:55.042662 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 15:36:55.044777 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 15:36:55.063996 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 15:36:55.065429 systemd-modules-load[180]: Inserted module 'br_netfilter' Feb 13 15:36:55.250215 kernel: Bridge firewalling registered Feb 13 15:36:55.066388 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 15:36:55.265608 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 15:36:55.270597 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 15:36:55.273103 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 15:36:55.281321 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 15:36:55.292503 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 15:36:55.293047 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 15:36:55.293970 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 15:36:55.299409 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 15:36:55.322804 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 15:36:55.333028 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 15:36:55.336467 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 15:36:55.344382 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 15:36:55.361375 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 15:36:55.414807 dracut-cmdline[213]: dracut-dracut-053 Feb 13 15:36:55.424088 dracut-cmdline[213]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=cd73eba291b8356dfc2c39f651cabef9206685f772c8949188fd366788d672c2 Feb 13 15:36:55.460762 systemd-resolved[214]: Positive Trust Anchors: Feb 13 15:36:55.460780 systemd-resolved[214]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 15:36:55.460923 systemd-resolved[214]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 15:36:55.479373 systemd-resolved[214]: Defaulting to hostname 'linux'. Feb 13 15:36:55.482581 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 15:36:55.485934 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 15:36:55.533212 kernel: SCSI subsystem initialized Feb 13 15:36:55.543148 kernel: Loading iSCSI transport class v2.0-870. Feb 13 15:36:55.558151 kernel: iscsi: registered transport (tcp) Feb 13 15:36:55.584444 kernel: iscsi: registered transport (qla4xxx) Feb 13 15:36:55.584526 kernel: QLogic iSCSI HBA Driver Feb 13 15:36:55.631744 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 15:36:55.639361 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 15:36:55.670153 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 15:36:55.670233 kernel: device-mapper: uevent: version 1.0.3 Feb 13 15:36:55.670266 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 15:36:55.716163 kernel: raid6: avx512x4 gen() 15414 MB/s Feb 13 15:36:55.733163 kernel: raid6: avx512x2 gen() 14100 MB/s Feb 13 15:36:55.750158 kernel: raid6: avx512x1 gen() 15707 MB/s Feb 13 15:36:55.768160 kernel: raid6: avx2x4 gen() 14995 MB/s Feb 13 15:36:55.786180 kernel: raid6: avx2x2 gen() 12114 MB/s Feb 13 15:36:55.804064 kernel: raid6: avx2x1 gen() 6597 MB/s Feb 13 15:36:55.804185 kernel: raid6: using algorithm avx512x1 gen() 15707 MB/s Feb 13 15:36:55.821153 kernel: raid6: .... xor() 13863 MB/s, rmw enabled Feb 13 15:36:55.821276 kernel: raid6: using avx512x2 recovery algorithm Feb 13 15:36:55.849159 kernel: xor: automatically using best checksumming function avx Feb 13 15:36:56.050161 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 15:36:56.062858 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 15:36:56.072589 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 15:36:56.103034 systemd-udevd[397]: Using default interface naming scheme 'v255'. Feb 13 15:36:56.117530 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 15:36:56.129302 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 15:36:56.166569 dracut-pre-trigger[405]: rd.md=0: removing MD RAID activation Feb 13 15:36:56.217564 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 15:36:56.222420 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 15:36:56.301592 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 15:36:56.317488 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 15:36:56.361105 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 15:36:56.364320 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 15:36:56.369667 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 15:36:56.371346 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 15:36:56.378898 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 15:36:56.425344 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 15:36:56.446024 kernel: ena 0000:00:05.0: ENA device version: 0.10 Feb 13 15:36:56.470247 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Feb 13 15:36:56.470970 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Feb 13 15:36:56.471339 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem febf4000, mac addr 06:ac:2d:85:ad:cf Feb 13 15:36:56.471647 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 15:36:56.472759 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 15:36:56.474725 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 15:36:56.478556 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 15:36:56.480324 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 15:36:56.489315 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 15:36:56.489358 kernel: AES CTR mode by8 optimization enabled Feb 13 15:36:56.480533 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 15:36:56.481975 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 15:36:56.499365 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 15:36:56.502504 (udev-worker)[457]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:36:56.597406 kernel: nvme nvme0: pci function 0000:00:04.0 Feb 13 15:36:56.597877 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Feb 13 15:36:56.609684 kernel: nvme nvme0: 2/0/0 default/read/poll queues Feb 13 15:36:56.618053 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 15:36:56.618146 kernel: GPT:9289727 != 16777215 Feb 13 15:36:56.618185 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 15:36:56.618206 kernel: GPT:9289727 != 16777215 Feb 13 15:36:56.618226 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 15:36:56.618244 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 15:36:56.730222 kernel: BTRFS: device fsid 966d6124-9067-4089-b000-5e99065fe7e2 devid 1 transid 37 /dev/nvme0n1p3 scanned by (udev-worker) (457) Feb 13 15:36:56.790171 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by (udev-worker) (456) Feb 13 15:36:56.819082 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 15:36:56.830890 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 15:36:56.855938 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Feb 13 15:36:56.862844 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Feb 13 15:36:56.897390 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Feb 13 15:36:56.901314 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 15:36:56.918734 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Feb 13 15:36:56.936899 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 15:36:56.944703 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 15:36:56.963286 disk-uuid[628]: Primary Header is updated. Feb 13 15:36:56.963286 disk-uuid[628]: Secondary Entries is updated. Feb 13 15:36:56.963286 disk-uuid[628]: Secondary Header is updated. Feb 13 15:36:56.972157 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 15:36:56.989252 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 15:36:58.002342 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 15:36:58.002420 disk-uuid[629]: The operation has completed successfully. Feb 13 15:36:58.245874 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 15:36:58.246003 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 15:36:58.267391 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 15:36:58.272845 sh[887]: Success Feb 13 15:36:58.288149 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 13 15:36:58.395423 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 15:36:58.420531 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 15:36:58.445616 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 15:36:58.487796 kernel: BTRFS info (device dm-0): first mount of filesystem 966d6124-9067-4089-b000-5e99065fe7e2 Feb 13 15:36:58.487922 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 15:36:58.487947 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 15:36:58.492717 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 15:36:58.492790 kernel: BTRFS info (device dm-0): using free space tree Feb 13 15:36:58.584157 kernel: BTRFS info (device dm-0): enabling ssd optimizations Feb 13 15:36:58.596119 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 15:36:58.610158 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 15:36:58.632770 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 15:36:58.647670 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 15:36:58.693702 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 83f602a1-06be-4b8b-b461-5e4f70db8da1 Feb 13 15:36:58.693788 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 15:36:58.693811 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 15:36:58.707168 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 15:36:58.721712 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 15:36:58.725242 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 83f602a1-06be-4b8b-b461-5e4f70db8da1 Feb 13 15:36:58.732636 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 15:36:58.746309 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 15:36:58.854287 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 15:36:58.867455 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 15:36:58.935499 systemd-networkd[1080]: lo: Link UP Feb 13 15:36:58.938116 systemd-networkd[1080]: lo: Gained carrier Feb 13 15:36:58.941440 systemd-networkd[1080]: Enumeration completed Feb 13 15:36:58.943325 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 15:36:58.943822 systemd-networkd[1080]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 15:36:58.943827 systemd-networkd[1080]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 15:36:58.949506 systemd[1]: Reached target network.target - Network. Feb 13 15:36:58.959031 systemd-networkd[1080]: eth0: Link UP Feb 13 15:36:58.959209 systemd-networkd[1080]: eth0: Gained carrier Feb 13 15:36:58.959229 systemd-networkd[1080]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 15:36:58.980321 systemd-networkd[1080]: eth0: DHCPv4 address 172.31.20.193/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 15:36:59.097972 ignition[994]: Ignition 2.20.0 Feb 13 15:36:59.097984 ignition[994]: Stage: fetch-offline Feb 13 15:36:59.098400 ignition[994]: no configs at "/usr/lib/ignition/base.d" Feb 13 15:36:59.098412 ignition[994]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:36:59.100548 ignition[994]: Ignition finished successfully Feb 13 15:36:59.105579 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 15:36:59.115458 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 15:36:59.173015 ignition[1090]: Ignition 2.20.0 Feb 13 15:36:59.173030 ignition[1090]: Stage: fetch Feb 13 15:36:59.173742 ignition[1090]: no configs at "/usr/lib/ignition/base.d" Feb 13 15:36:59.173757 ignition[1090]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:36:59.173870 ignition[1090]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:36:59.236150 ignition[1090]: PUT result: OK Feb 13 15:36:59.261155 ignition[1090]: parsed url from cmdline: "" Feb 13 15:36:59.261168 ignition[1090]: no config URL provided Feb 13 15:36:59.261178 ignition[1090]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 15:36:59.261218 ignition[1090]: no config at "/usr/lib/ignition/user.ign" Feb 13 15:36:59.261246 ignition[1090]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:36:59.262311 ignition[1090]: PUT result: OK Feb 13 15:36:59.262361 ignition[1090]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Feb 13 15:36:59.269317 ignition[1090]: GET result: OK Feb 13 15:36:59.269378 ignition[1090]: parsing config with SHA512: 56c9192811c267e34a4db80d7e9bee375ca6eda85091dc0beb1b37164b9fb2d21b6215883ebae77bf9898a405012548c2cdae8b7152f5592adc26a4afeba0bc0 Feb 13 15:36:59.279019 unknown[1090]: fetched base config from "system" Feb 13 15:36:59.279033 unknown[1090]: fetched base config from "system" Feb 13 15:36:59.279684 ignition[1090]: fetch: fetch complete Feb 13 15:36:59.279040 unknown[1090]: fetched user config from "aws" Feb 13 15:36:59.279691 ignition[1090]: fetch: fetch passed Feb 13 15:36:59.283729 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 15:36:59.279752 ignition[1090]: Ignition finished successfully Feb 13 15:36:59.304357 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 15:36:59.350764 ignition[1097]: Ignition 2.20.0 Feb 13 15:36:59.350778 ignition[1097]: Stage: kargs Feb 13 15:36:59.351523 ignition[1097]: no configs at "/usr/lib/ignition/base.d" Feb 13 15:36:59.351537 ignition[1097]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:36:59.351653 ignition[1097]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:36:59.356451 ignition[1097]: PUT result: OK Feb 13 15:36:59.375699 ignition[1097]: kargs: kargs passed Feb 13 15:36:59.376390 ignition[1097]: Ignition finished successfully Feb 13 15:36:59.378246 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 15:36:59.397160 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 15:36:59.481228 ignition[1104]: Ignition 2.20.0 Feb 13 15:36:59.481266 ignition[1104]: Stage: disks Feb 13 15:36:59.482576 ignition[1104]: no configs at "/usr/lib/ignition/base.d" Feb 13 15:36:59.482591 ignition[1104]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:36:59.484197 ignition[1104]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:36:59.491024 ignition[1104]: PUT result: OK Feb 13 15:36:59.501511 ignition[1104]: disks: disks passed Feb 13 15:36:59.501610 ignition[1104]: Ignition finished successfully Feb 13 15:36:59.508110 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 15:36:59.516030 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 15:36:59.518976 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 15:36:59.526464 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 15:36:59.532233 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 15:36:59.540114 systemd[1]: Reached target basic.target - Basic System. Feb 13 15:36:59.552474 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 15:36:59.631253 systemd-fsck[1113]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 15:36:59.637594 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 15:36:59.662711 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 15:36:59.858908 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 85ed0b0d-7f0f-4eeb-80d8-6213e9fcc55d r/w with ordered data mode. Quota mode: none. Feb 13 15:36:59.858823 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 15:36:59.861980 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 15:36:59.872442 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 15:36:59.883296 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 15:36:59.888931 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 15:36:59.889014 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 15:36:59.889053 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 15:36:59.915726 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1132) Feb 13 15:36:59.918454 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 83f602a1-06be-4b8b-b461-5e4f70db8da1 Feb 13 15:36:59.918503 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 15:36:59.918517 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 15:36:59.923233 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 15:36:59.928148 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 15:36:59.929410 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 15:36:59.938285 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 15:37:00.187279 initrd-setup-root[1156]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 15:37:00.207643 initrd-setup-root[1163]: cut: /sysroot/etc/group: No such file or directory Feb 13 15:37:00.214940 initrd-setup-root[1170]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 15:37:00.240910 initrd-setup-root[1177]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 15:37:00.501526 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 15:37:00.511317 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 15:37:00.532988 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 15:37:00.556891 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 83f602a1-06be-4b8b-b461-5e4f70db8da1 Feb 13 15:37:00.561246 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 15:37:00.642538 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 15:37:00.656784 ignition[1245]: INFO : Ignition 2.20.0 Feb 13 15:37:00.656784 ignition[1245]: INFO : Stage: mount Feb 13 15:37:00.662729 ignition[1245]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 15:37:00.662729 ignition[1245]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:37:00.662729 ignition[1245]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:37:00.662729 ignition[1245]: INFO : PUT result: OK Feb 13 15:37:00.679121 ignition[1245]: INFO : mount: mount passed Feb 13 15:37:00.687541 ignition[1245]: INFO : Ignition finished successfully Feb 13 15:37:00.693786 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 15:37:00.708761 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 15:37:00.863414 systemd-networkd[1080]: eth0: Gained IPv6LL Feb 13 15:37:00.871447 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 15:37:00.917161 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by mount (1256) Feb 13 15:37:00.919628 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 83f602a1-06be-4b8b-b461-5e4f70db8da1 Feb 13 15:37:00.919698 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 15:37:00.919718 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 15:37:00.930197 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 15:37:00.939162 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 15:37:01.000034 ignition[1273]: INFO : Ignition 2.20.0 Feb 13 15:37:01.000034 ignition[1273]: INFO : Stage: files Feb 13 15:37:01.005721 ignition[1273]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 15:37:01.005721 ignition[1273]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:37:01.005721 ignition[1273]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:37:01.020690 ignition[1273]: INFO : PUT result: OK Feb 13 15:37:01.031347 ignition[1273]: DEBUG : files: compiled without relabeling support, skipping Feb 13 15:37:01.036345 ignition[1273]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 15:37:01.036345 ignition[1273]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 15:37:01.066356 ignition[1273]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 15:37:01.068075 ignition[1273]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 15:37:01.068075 ignition[1273]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 15:37:01.067108 unknown[1273]: wrote ssh authorized keys file for user: core Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Feb 13 15:37:01.081738 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Feb 13 15:37:01.527160 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Feb 13 15:37:02.139251 ignition[1273]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Feb 13 15:37:02.139251 ignition[1273]: INFO : files: op(8): [started] processing unit "containerd.service" Feb 13 15:37:02.149306 ignition[1273]: INFO : files: op(8): op(9): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Feb 13 15:37:02.153693 ignition[1273]: INFO : files: op(8): op(9): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Feb 13 15:37:02.153693 ignition[1273]: INFO : files: op(8): [finished] processing unit "containerd.service" Feb 13 15:37:02.163406 ignition[1273]: INFO : files: createResultFile: createFiles: op(a): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 15:37:02.163406 ignition[1273]: INFO : files: createResultFile: createFiles: op(a): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 15:37:02.163406 ignition[1273]: INFO : files: files passed Feb 13 15:37:02.163406 ignition[1273]: INFO : Ignition finished successfully Feb 13 15:37:02.173679 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 15:37:02.191392 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 15:37:02.205931 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 15:37:02.217699 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 15:37:02.217841 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 15:37:02.229156 initrd-setup-root-after-ignition[1302]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 15:37:02.229156 initrd-setup-root-after-ignition[1302]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 15:37:02.236027 initrd-setup-root-after-ignition[1306]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 15:37:02.245298 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 15:37:02.250663 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 15:37:02.256454 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 15:37:02.322383 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 15:37:02.322509 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 15:37:02.329534 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 15:37:02.333564 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 15:37:02.336534 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 15:37:02.345361 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 15:37:02.365657 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 15:37:02.379898 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 15:37:02.416045 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 15:37:02.420471 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 15:37:02.424270 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 15:37:02.427373 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 15:37:02.427548 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 15:37:02.432949 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 15:37:02.436619 systemd[1]: Stopped target basic.target - Basic System. Feb 13 15:37:02.439464 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 15:37:02.442024 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 15:37:02.442197 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 15:37:02.447682 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 15:37:02.450840 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 15:37:02.452547 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 15:37:02.454028 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 15:37:02.459401 systemd[1]: Stopped target swap.target - Swaps. Feb 13 15:37:02.460519 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 15:37:02.460658 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 15:37:02.467187 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 15:37:02.470404 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 15:37:02.477625 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 15:37:02.479240 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 15:37:02.486936 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 15:37:02.489061 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 15:37:02.492333 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 15:37:02.492580 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 15:37:02.498910 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 15:37:02.499030 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 15:37:02.519304 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 15:37:02.549306 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 15:37:02.553089 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 15:37:02.556582 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 15:37:02.563710 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 15:37:02.564149 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 15:37:02.578655 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 15:37:02.578789 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 15:37:02.602282 ignition[1326]: INFO : Ignition 2.20.0 Feb 13 15:37:02.602282 ignition[1326]: INFO : Stage: umount Feb 13 15:37:02.602282 ignition[1326]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 15:37:02.602282 ignition[1326]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 15:37:02.602282 ignition[1326]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 15:37:02.602282 ignition[1326]: INFO : PUT result: OK Feb 13 15:37:02.618821 ignition[1326]: INFO : umount: umount passed Feb 13 15:37:02.618821 ignition[1326]: INFO : Ignition finished successfully Feb 13 15:37:02.618107 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 15:37:02.618272 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 15:37:02.620975 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 15:37:02.621028 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 15:37:02.625549 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 15:37:02.625645 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 15:37:02.630310 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 15:37:02.630390 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 15:37:02.639807 systemd[1]: Stopped target network.target - Network. Feb 13 15:37:02.646617 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 15:37:02.651111 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 15:37:02.655296 systemd[1]: Stopped target paths.target - Path Units. Feb 13 15:37:02.663570 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 15:37:02.671178 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 15:37:02.674147 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 15:37:02.678817 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 15:37:02.680308 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 15:37:02.680372 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 15:37:02.683277 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 15:37:02.683338 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 15:37:02.686452 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 15:37:02.688045 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 15:37:02.691024 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 15:37:02.691115 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 15:37:02.695740 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 15:37:02.698369 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 15:37:02.703624 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 15:37:02.704167 systemd-networkd[1080]: eth0: DHCPv6 lease lost Feb 13 15:37:02.704579 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 15:37:02.704700 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 15:37:02.721172 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 15:37:02.721334 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 15:37:02.726038 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 15:37:02.726161 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 15:37:02.739330 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 15:37:02.739411 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 15:37:02.741537 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 15:37:02.742884 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 15:37:02.756311 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 15:37:02.756433 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 15:37:02.762254 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 15:37:02.764277 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 15:37:02.764338 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 15:37:02.765743 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 15:37:02.765806 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 15:37:02.768866 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 15:37:02.768999 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 15:37:02.771849 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 15:37:02.799997 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 15:37:02.800326 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 15:37:02.804678 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 15:37:02.804726 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 15:37:02.809139 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 15:37:02.809272 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 15:37:02.811231 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 15:37:02.811292 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 15:37:02.814026 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 15:37:02.814084 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 15:37:02.817030 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 15:37:02.818703 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 15:37:02.846402 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 15:37:02.852155 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 15:37:02.852498 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 15:37:02.862625 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Feb 13 15:37:02.862709 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 15:37:02.866952 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 15:37:02.867014 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 15:37:02.869368 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 15:37:02.869436 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 15:37:02.874536 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 15:37:02.874644 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 15:37:02.901665 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 15:37:02.902172 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 15:37:02.903501 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 15:37:02.913800 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 15:37:02.983001 systemd[1]: Switching root. Feb 13 15:37:03.023849 systemd-journald[179]: Journal stopped Feb 13 15:37:05.313229 systemd-journald[179]: Received SIGTERM from PID 1 (systemd). Feb 13 15:37:05.313321 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 15:37:05.313354 kernel: SELinux: policy capability open_perms=1 Feb 13 15:37:05.313373 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 15:37:05.313392 kernel: SELinux: policy capability always_check_network=0 Feb 13 15:37:05.313410 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 15:37:05.313431 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 15:37:05.313459 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 15:37:05.313478 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 15:37:05.313496 kernel: audit: type=1403 audit(1739461023.691:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 15:37:05.313517 systemd[1]: Successfully loaded SELinux policy in 57.508ms. Feb 13 15:37:05.313546 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.176ms. Feb 13 15:37:05.313569 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 15:37:05.313591 systemd[1]: Detected virtualization amazon. Feb 13 15:37:05.313614 systemd[1]: Detected architecture x86-64. Feb 13 15:37:05.313641 systemd[1]: Detected first boot. Feb 13 15:37:05.313662 systemd[1]: Initializing machine ID from VM UUID. Feb 13 15:37:05.313685 zram_generator::config[1385]: No configuration found. Feb 13 15:37:05.313712 systemd[1]: Populated /etc with preset unit settings. Feb 13 15:37:05.313735 systemd[1]: Queued start job for default target multi-user.target. Feb 13 15:37:05.313763 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Feb 13 15:37:05.313787 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 15:37:05.313811 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 15:37:05.313839 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 15:37:05.313868 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 15:37:05.313891 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 15:37:05.313914 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 15:37:05.313937 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 15:37:05.313960 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 15:37:05.313983 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 15:37:05.314005 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 15:37:05.314028 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 15:37:05.314055 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 15:37:05.314078 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 15:37:05.314100 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 15:37:05.314186 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 15:37:05.314208 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 15:37:05.314232 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 15:37:05.314251 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 15:37:05.314269 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 15:37:05.314288 systemd[1]: Reached target slices.target - Slice Units. Feb 13 15:37:05.315660 systemd[1]: Reached target swap.target - Swaps. Feb 13 15:37:05.315690 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 15:37:05.315715 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 15:37:05.315737 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 15:37:05.315761 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 15:37:05.315784 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 15:37:05.315806 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 15:37:05.316021 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 15:37:05.316058 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 15:37:05.316081 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 15:37:05.316103 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 15:37:05.316138 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 15:37:05.316157 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:05.316176 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 15:37:05.316194 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 15:37:05.316212 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 15:37:05.316231 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 15:37:05.316253 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 15:37:05.316272 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 15:37:05.316290 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 15:37:05.316308 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 15:37:05.316326 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 15:37:05.316345 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 15:37:05.316363 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 15:37:05.316382 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 15:37:05.316408 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 15:37:05.316431 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Feb 13 15:37:05.316453 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Feb 13 15:37:05.316471 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 15:37:05.316490 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 15:37:05.316507 kernel: fuse: init (API version 7.39) Feb 13 15:37:05.316528 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 15:37:05.316547 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 15:37:05.316566 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 15:37:05.316590 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:05.316609 kernel: loop: module loaded Feb 13 15:37:05.316667 systemd-journald[1482]: Collecting audit messages is disabled. Feb 13 15:37:05.316703 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 15:37:05.316722 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 15:37:05.316741 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 15:37:05.316820 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 15:37:05.316845 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 15:37:05.316863 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 15:37:05.316885 systemd-journald[1482]: Journal started Feb 13 15:37:05.316923 systemd-journald[1482]: Runtime Journal (/run/log/journal/ec2668ca7830d0977d2a9a726ac7b55a) is 4.8M, max 38.6M, 33.7M free. Feb 13 15:37:05.333248 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 15:37:05.333342 kernel: ACPI: bus type drm_connector registered Feb 13 15:37:05.328774 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 15:37:05.331585 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 15:37:05.332232 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 15:37:05.335150 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 15:37:05.335536 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 15:37:05.337981 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 15:37:05.338403 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 15:37:05.342043 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 15:37:05.342865 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 15:37:05.345548 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 15:37:05.345857 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 15:37:05.348077 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 15:37:05.350806 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 15:37:05.355413 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 15:37:05.370824 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 15:37:05.376473 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 15:37:05.387454 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 15:37:05.403063 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 15:37:05.411360 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 15:37:05.424337 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 15:37:05.426668 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 15:37:05.437476 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 15:37:05.455497 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 15:37:05.457450 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 15:37:05.461990 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 15:37:05.464112 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 15:37:05.477390 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 15:37:05.494339 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 15:37:05.504319 systemd-journald[1482]: Time spent on flushing to /var/log/journal/ec2668ca7830d0977d2a9a726ac7b55a is 77.957ms for 930 entries. Feb 13 15:37:05.504319 systemd-journald[1482]: System Journal (/var/log/journal/ec2668ca7830d0977d2a9a726ac7b55a) is 8.0M, max 195.6M, 187.6M free. Feb 13 15:37:05.607434 systemd-journald[1482]: Received client request to flush runtime journal. Feb 13 15:37:05.508468 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 15:37:05.516312 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 15:37:05.522754 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 15:37:05.539443 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 15:37:05.542154 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 15:37:05.546185 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 15:37:05.580572 udevadm[1540]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Feb 13 15:37:05.612772 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 15:37:05.622896 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 15:37:05.634557 systemd-tmpfiles[1535]: ACLs are not supported, ignoring. Feb 13 15:37:05.634585 systemd-tmpfiles[1535]: ACLs are not supported, ignoring. Feb 13 15:37:05.645591 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 15:37:05.662350 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 15:37:05.733721 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 15:37:05.742424 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 15:37:05.796853 systemd-tmpfiles[1557]: ACLs are not supported, ignoring. Feb 13 15:37:05.797349 systemd-tmpfiles[1557]: ACLs are not supported, ignoring. Feb 13 15:37:05.813403 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 15:37:06.500899 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 15:37:06.509562 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 15:37:06.553967 systemd-udevd[1563]: Using default interface naming scheme 'v255'. Feb 13 15:37:06.615693 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 15:37:06.626658 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 15:37:06.666552 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 15:37:06.743641 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Feb 13 15:37:06.783736 (udev-worker)[1571]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:37:06.802520 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 15:37:06.850557 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Feb 13 15:37:06.922191 kernel: ACPI: button: Power Button [PWRF] Feb 13 15:37:06.927627 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Feb 13 15:37:06.943152 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0xb100, revision 255 Feb 13 15:37:06.970001 kernel: ACPI: button: Sleep Button [SLPF] Feb 13 15:37:06.951047 systemd-networkd[1568]: lo: Link UP Feb 13 15:37:06.995717 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input5 Feb 13 15:37:06.951053 systemd-networkd[1568]: lo: Gained carrier Feb 13 15:37:06.953294 systemd-networkd[1568]: Enumeration completed Feb 13 15:37:06.953595 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 15:37:06.954156 systemd-networkd[1568]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 15:37:06.954162 systemd-networkd[1568]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 15:37:06.957105 systemd-networkd[1568]: eth0: Link UP Feb 13 15:37:06.957502 systemd-networkd[1568]: eth0: Gained carrier Feb 13 15:37:06.957533 systemd-networkd[1568]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 15:37:06.963443 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 15:37:06.972216 systemd-networkd[1568]: eth0: DHCPv4 address 172.31.20.193/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 15:37:06.983379 systemd-networkd[1568]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 15:37:07.015950 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 15:37:07.043165 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1564) Feb 13 15:37:07.056957 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 15:37:07.262202 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 15:37:07.420336 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 15:37:07.429991 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 15:37:07.449074 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 15:37:07.474621 lvm[1687]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 15:37:07.507522 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 15:37:07.510726 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 15:37:07.520581 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 15:37:07.532385 lvm[1690]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 15:37:07.570536 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 15:37:07.578078 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 15:37:07.581657 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 15:37:07.581693 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 15:37:07.585739 systemd[1]: Reached target machines.target - Containers. Feb 13 15:37:07.588518 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Feb 13 15:37:07.599537 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 15:37:07.613943 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 15:37:07.626921 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 15:37:07.630013 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 15:37:07.653252 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Feb 13 15:37:07.676590 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 15:37:07.686901 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 15:37:07.715059 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 15:37:07.743168 kernel: loop0: detected capacity change from 0 to 211296 Feb 13 15:37:07.756551 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 15:37:07.769748 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Feb 13 15:37:07.874149 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 15:37:07.899209 kernel: loop1: detected capacity change from 0 to 62848 Feb 13 15:37:07.954191 kernel: loop2: detected capacity change from 0 to 138184 Feb 13 15:37:08.048139 kernel: loop3: detected capacity change from 0 to 140992 Feb 13 15:37:08.183164 kernel: loop4: detected capacity change from 0 to 211296 Feb 13 15:37:08.229165 kernel: loop5: detected capacity change from 0 to 62848 Feb 13 15:37:08.251294 kernel: loop6: detected capacity change from 0 to 138184 Feb 13 15:37:08.287404 systemd-networkd[1568]: eth0: Gained IPv6LL Feb 13 15:37:08.294321 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 15:37:08.304992 kernel: loop7: detected capacity change from 0 to 140992 Feb 13 15:37:08.348637 (sd-merge)[1712]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Feb 13 15:37:08.349803 (sd-merge)[1712]: Merged extensions into '/usr'. Feb 13 15:37:08.366013 systemd[1]: Reloading requested from client PID 1698 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 15:37:08.366102 systemd[1]: Reloading... Feb 13 15:37:08.548334 zram_generator::config[1741]: No configuration found. Feb 13 15:37:08.775886 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 15:37:08.879367 systemd[1]: Reloading finished in 510 ms. Feb 13 15:37:08.900028 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 15:37:08.913483 systemd[1]: Starting ensure-sysext.service... Feb 13 15:37:08.922357 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 15:37:08.940399 systemd[1]: Reloading requested from client PID 1795 ('systemctl') (unit ensure-sysext.service)... Feb 13 15:37:08.940421 systemd[1]: Reloading... Feb 13 15:37:08.957392 systemd-tmpfiles[1796]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 15:37:08.964691 systemd-tmpfiles[1796]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 15:37:08.971493 systemd-tmpfiles[1796]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 15:37:08.972057 systemd-tmpfiles[1796]: ACLs are not supported, ignoring. Feb 13 15:37:08.972210 systemd-tmpfiles[1796]: ACLs are not supported, ignoring. Feb 13 15:37:08.995320 systemd-tmpfiles[1796]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 15:37:08.995392 systemd-tmpfiles[1796]: Skipping /boot Feb 13 15:37:09.012247 systemd-tmpfiles[1796]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 15:37:09.012262 systemd-tmpfiles[1796]: Skipping /boot Feb 13 15:37:09.126236 ldconfig[1694]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 15:37:09.134230 zram_generator::config[1824]: No configuration found. Feb 13 15:37:09.444791 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 15:37:09.551163 systemd[1]: Reloading finished in 609 ms. Feb 13 15:37:09.569012 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 15:37:09.581535 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 15:37:09.593562 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 15:37:09.606687 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 15:37:09.616346 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 15:37:09.633712 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 15:37:09.643365 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 15:37:09.716365 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.717381 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 15:37:09.727571 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 15:37:09.743195 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 15:37:09.771668 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 15:37:09.773289 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 15:37:09.773806 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.790629 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.792251 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 15:37:09.792599 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 15:37:09.792823 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.806257 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 15:37:09.806497 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 15:37:09.813025 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.815567 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 15:37:09.831529 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 15:37:09.833228 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 15:37:09.833443 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 15:37:09.833679 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 15:37:09.838492 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 15:37:09.849106 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 15:37:09.858684 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 15:37:09.861592 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 15:37:09.861966 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 15:37:09.866095 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 15:37:09.866703 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 15:37:09.877618 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 15:37:09.878005 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 15:37:09.902259 systemd[1]: Finished ensure-sysext.service. Feb 13 15:37:09.904813 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 15:37:09.916436 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 15:37:09.923858 augenrules[1929]: No rules Feb 13 15:37:09.925505 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 15:37:09.932385 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 15:37:09.954120 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 15:37:09.965944 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 15:37:09.973969 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 15:37:09.981509 systemd-resolved[1887]: Positive Trust Anchors: Feb 13 15:37:09.981530 systemd-resolved[1887]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 15:37:09.981579 systemd-resolved[1887]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 15:37:09.990861 systemd-resolved[1887]: Defaulting to hostname 'linux'. Feb 13 15:37:10.005452 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 15:37:10.009074 systemd[1]: Reached target network.target - Network. Feb 13 15:37:10.010371 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 15:37:10.013019 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 15:37:10.015633 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 15:37:10.017494 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 15:37:10.019194 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 15:37:10.020818 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 15:37:10.022195 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 15:37:10.023624 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 15:37:10.025331 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 15:37:10.025367 systemd[1]: Reached target paths.target - Path Units. Feb 13 15:37:10.026616 systemd[1]: Reached target timers.target - Timer Units. Feb 13 15:37:10.029517 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 15:37:10.035148 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 15:37:10.038980 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 15:37:10.042550 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 15:37:10.044845 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 15:37:10.046615 systemd[1]: Reached target basic.target - Basic System. Feb 13 15:37:10.048460 systemd[1]: System is tainted: cgroupsv1 Feb 13 15:37:10.048522 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 15:37:10.048554 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 15:37:10.057441 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 15:37:10.064585 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 15:37:10.077407 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 15:37:10.090357 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 15:37:10.103634 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 15:37:10.113449 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 15:37:10.150390 jq[1946]: false Feb 13 15:37:10.152812 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 15:37:10.176613 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 15:37:10.195355 systemd[1]: Started ntpd.service - Network Time Service. Feb 13 15:37:10.203378 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 15:37:10.222356 systemd[1]: Starting setup-oem.service - Setup OEM... Feb 13 15:37:10.234742 extend-filesystems[1947]: Found loop4 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found loop5 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found loop6 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found loop7 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p1 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p2 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p3 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found usr Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p4 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p6 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p7 Feb 13 15:37:10.237204 extend-filesystems[1947]: Found nvme0n1p9 Feb 13 15:37:10.237204 extend-filesystems[1947]: Checking size of /dev/nvme0n1p9 Feb 13 15:37:10.243048 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 15:37:10.259443 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 15:37:10.274786 dbus-daemon[1944]: [system] SELinux support is enabled Feb 13 15:37:10.282488 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 15:37:10.285655 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 15:37:10.298064 dbus-daemon[1944]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1568 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Feb 13 15:37:10.310679 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 15:37:10.324918 extend-filesystems[1947]: Resized partition /dev/nvme0n1p9 Feb 13 15:37:10.330264 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 15:37:10.343158 extend-filesystems[1975]: resize2fs 1.47.1 (20-May-2024) Feb 13 15:37:10.346602 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 15:37:10.361166 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Feb 13 15:37:10.377915 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 15:37:10.378559 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 15:37:10.408362 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 15:37:10.408866 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 15:37:10.441318 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 15:37:10.441788 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 15:37:10.445096 ntpd[1955]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 13:21:05 UTC 2025 (1): Starting Feb 13 15:37:10.446580 ntpd[1955]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 13:21:05 UTC 2025 (1): Starting Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: ---------------------------------------------------- Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: ntp-4 is maintained by Network Time Foundation, Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: corporation. Support and training for ntp-4 are Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: available at https://www.nwtime.org/support Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: ---------------------------------------------------- Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: proto: precision = 0.086 usec (-23) Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: basedate set to 2025-02-01 Feb 13 15:37:10.456693 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: gps base set to 2025-02-02 (week 2352) Feb 13 15:37:10.446595 ntpd[1955]: ---------------------------------------------------- Feb 13 15:37:10.563816 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen normally on 3 eth0 172.31.20.193:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen normally on 4 lo [::1]:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listen normally on 5 eth0 [fe80::4ac:2dff:fe85:adcf%2]:123 Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: Listening on routing socket on fd #22 for interface updates Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 15:37:10.564098 ntpd[1955]: 13 Feb 15:37:10 ntpd[1955]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 15:37:10.564546 jq[1974]: true Feb 13 15:37:10.523451 (ntainerd)[1992]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 15:37:10.446607 ntpd[1955]: ntp-4 is maintained by Network Time Foundation, Feb 13 15:37:10.446616 ntpd[1955]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 15:37:10.446628 ntpd[1955]: corporation. Support and training for ntp-4 are Feb 13 15:37:10.446639 ntpd[1955]: available at https://www.nwtime.org/support Feb 13 15:37:10.446649 ntpd[1955]: ---------------------------------------------------- Feb 13 15:37:10.453699 ntpd[1955]: proto: precision = 0.086 usec (-23) Feb 13 15:37:10.455217 ntpd[1955]: basedate set to 2025-02-01 Feb 13 15:37:10.455240 ntpd[1955]: gps base set to 2025-02-02 (week 2352) Feb 13 15:37:10.463795 ntpd[1955]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 15:37:10.463856 ntpd[1955]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 15:37:10.464498 ntpd[1955]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 15:37:10.464542 ntpd[1955]: Listen normally on 3 eth0 172.31.20.193:123 Feb 13 15:37:10.464583 ntpd[1955]: Listen normally on 4 lo [::1]:123 Feb 13 15:37:10.464626 ntpd[1955]: Listen normally on 5 eth0 [fe80::4ac:2dff:fe85:adcf%2]:123 Feb 13 15:37:10.464664 ntpd[1955]: Listening on routing socket on fd #22 for interface updates Feb 13 15:37:10.510244 ntpd[1955]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 15:37:10.510282 ntpd[1955]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 15:37:10.571459 extend-filesystems[1975]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Feb 13 15:37:10.571459 extend-filesystems[1975]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 15:37:10.571459 extend-filesystems[1975]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Feb 13 15:37:10.577961 extend-filesystems[1947]: Resized filesystem in /dev/nvme0n1p9 Feb 13 15:37:10.574636 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 15:37:10.575028 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 15:37:10.599408 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 15:37:10.626050 systemd[1]: Finished setup-oem.service - Setup OEM. Feb 13 15:37:10.629842 jq[2003]: true Feb 13 15:37:10.639810 dbus-daemon[1944]: [system] Successfully activated service 'org.freedesktop.systemd1' Feb 13 15:37:10.668687 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Feb 13 15:37:10.689995 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 15:37:10.749423 update_engine[1968]: I20250213 15:37:10.691633 1968 main.cc:92] Flatcar Update Engine starting Feb 13 15:37:10.690037 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 15:37:10.692887 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 15:37:10.692916 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 15:37:10.766742 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Feb 13 15:37:10.774208 update_engine[1968]: I20250213 15:37:10.772377 1968 update_check_scheduler.cc:74] Next update check in 11m39s Feb 13 15:37:10.777029 systemd[1]: Started update-engine.service - Update Engine. Feb 13 15:37:10.806690 coreos-metadata[1943]: Feb 13 15:37:10.803 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.808 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.815 INFO Fetch successful Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.815 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.820 INFO Fetch successful Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.820 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.825 INFO Fetch successful Feb 13 15:37:10.832632 coreos-metadata[1943]: Feb 13 15:37:10.825 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Feb 13 15:37:10.827018 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 15:37:10.830336 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 15:37:10.835427 coreos-metadata[1943]: Feb 13 15:37:10.834 INFO Fetch successful Feb 13 15:37:10.835427 coreos-metadata[1943]: Feb 13 15:37:10.834 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Feb 13 15:37:10.839939 systemd-logind[1965]: Watching system buttons on /dev/input/event1 (Power Button) Feb 13 15:37:10.854650 coreos-metadata[1943]: Feb 13 15:37:10.840 INFO Fetch failed with 404: resource not found Feb 13 15:37:10.854650 coreos-metadata[1943]: Feb 13 15:37:10.840 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Feb 13 15:37:10.860840 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (2033) Feb 13 15:37:10.860920 coreos-metadata[1943]: Feb 13 15:37:10.860 INFO Fetch successful Feb 13 15:37:10.860920 coreos-metadata[1943]: Feb 13 15:37:10.860 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Feb 13 15:37:10.855068 systemd-logind[1965]: Watching system buttons on /dev/input/event2 (Sleep Button) Feb 13 15:37:10.855173 systemd-logind[1965]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 13 15:37:10.861838 coreos-metadata[1943]: Feb 13 15:37:10.861 INFO Fetch successful Feb 13 15:37:10.861838 coreos-metadata[1943]: Feb 13 15:37:10.861 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Feb 13 15:37:10.863798 systemd-logind[1965]: New seat seat0. Feb 13 15:37:10.873737 coreos-metadata[1943]: Feb 13 15:37:10.867 INFO Fetch successful Feb 13 15:37:10.873737 coreos-metadata[1943]: Feb 13 15:37:10.867 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Feb 13 15:37:10.890244 coreos-metadata[1943]: Feb 13 15:37:10.880 INFO Fetch successful Feb 13 15:37:10.890244 coreos-metadata[1943]: Feb 13 15:37:10.880 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Feb 13 15:37:10.890244 coreos-metadata[1943]: Feb 13 15:37:10.881 INFO Fetch successful Feb 13 15:37:10.882856 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 15:37:11.020274 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 15:37:11.022280 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 15:37:11.087725 bash[2067]: Updated "/home/core/.ssh/authorized_keys" Feb 13 15:37:11.090281 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 15:37:11.137308 systemd[1]: Starting sshkeys.service... Feb 13 15:37:11.147906 amazon-ssm-agent[2027]: Initializing new seelog logger Feb 13 15:37:11.154393 amazon-ssm-agent[2027]: New Seelog Logger Creation Complete Feb 13 15:37:11.154393 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.154393 amazon-ssm-agent[2027]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.154393 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 processing appconfig overrides Feb 13 15:37:11.156497 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.163619 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO Proxy environment variables: Feb 13 15:37:11.165373 amazon-ssm-agent[2027]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.165513 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 processing appconfig overrides Feb 13 15:37:11.171471 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.171471 amazon-ssm-agent[2027]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.171471 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 processing appconfig overrides Feb 13 15:37:11.182477 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Feb 13 15:37:11.194240 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Feb 13 15:37:11.196958 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.197382 amazon-ssm-agent[2027]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 15:37:11.197382 amazon-ssm-agent[2027]: 2025/02/13 15:37:11 processing appconfig overrides Feb 13 15:37:11.223571 dbus-daemon[1944]: [system] Successfully activated service 'org.freedesktop.hostname1' Feb 13 15:37:11.227037 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Feb 13 15:37:11.231719 dbus-daemon[1944]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=2030 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Feb 13 15:37:11.257476 systemd[1]: Starting polkit.service - Authorization Manager... Feb 13 15:37:11.266091 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO https_proxy: Feb 13 15:37:11.308423 polkitd[2104]: Started polkitd version 121 Feb 13 15:37:11.346818 polkitd[2104]: Loading rules from directory /etc/polkit-1/rules.d Feb 13 15:37:11.346974 polkitd[2104]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 13 15:37:11.348082 polkitd[2104]: Finished loading, compiling and executing 2 rules Feb 13 15:37:11.349270 dbus-daemon[1944]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Feb 13 15:37:11.351795 systemd[1]: Started polkit.service - Authorization Manager. Feb 13 15:37:11.354690 polkitd[2104]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 13 15:37:11.377054 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO http_proxy: Feb 13 15:37:11.443636 systemd-resolved[1887]: System hostname changed to 'ip-172-31-20-193'. Feb 13 15:37:11.445596 systemd-hostnamed[2030]: Hostname set to (transient) Feb 13 15:37:11.467165 coreos-metadata[2094]: Feb 13 15:37:11.462 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 15:37:11.468460 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO no_proxy: Feb 13 15:37:11.474489 coreos-metadata[2094]: Feb 13 15:37:11.472 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Feb 13 15:37:11.475341 coreos-metadata[2094]: Feb 13 15:37:11.475 INFO Fetch successful Feb 13 15:37:11.475341 coreos-metadata[2094]: Feb 13 15:37:11.475 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Feb 13 15:37:11.482329 coreos-metadata[2094]: Feb 13 15:37:11.477 INFO Fetch successful Feb 13 15:37:11.484522 unknown[2094]: wrote ssh authorized keys file for user: core Feb 13 15:37:11.544747 update-ssh-keys[2169]: Updated "/home/core/.ssh/authorized_keys" Feb 13 15:37:11.549934 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Feb 13 15:37:11.568054 systemd[1]: Finished sshkeys.service. Feb 13 15:37:11.577183 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO Checking if agent identity type OnPrem can be assumed Feb 13 15:37:11.626395 locksmithd[2032]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 15:37:11.682884 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO Checking if agent identity type EC2 can be assumed Feb 13 15:37:11.789776 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO Agent will take identity from EC2 Feb 13 15:37:11.889600 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 15:37:11.980144 containerd[1992]: time="2025-02-13T15:37:11.977345281Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 15:37:11.987485 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 15:37:12.073390 containerd[1992]: time="2025-02-13T15:37:12.073275352Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.076078 containerd[1992]: time="2025-02-13T15:37:12.076033297Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 15:37:12.076223 containerd[1992]: time="2025-02-13T15:37:12.076206233Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 15:37:12.076300 containerd[1992]: time="2025-02-13T15:37:12.076286788Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 15:37:12.076536 containerd[1992]: time="2025-02-13T15:37:12.076520082Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 15:37:12.076608 containerd[1992]: time="2025-02-13T15:37:12.076596226Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.076743 containerd[1992]: time="2025-02-13T15:37:12.076727391Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 15:37:12.076805 containerd[1992]: time="2025-02-13T15:37:12.076794192Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077150 containerd[1992]: time="2025-02-13T15:37:12.077112492Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077242 containerd[1992]: time="2025-02-13T15:37:12.077220976Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077313 containerd[1992]: time="2025-02-13T15:37:12.077299794Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077372 containerd[1992]: time="2025-02-13T15:37:12.077360889Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077533 containerd[1992]: time="2025-02-13T15:37:12.077519968Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.077824 containerd[1992]: time="2025-02-13T15:37:12.077808656Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 15:37:12.078069 containerd[1992]: time="2025-02-13T15:37:12.078051070Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 15:37:12.078167 containerd[1992]: time="2025-02-13T15:37:12.078151502Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 15:37:12.078461 containerd[1992]: time="2025-02-13T15:37:12.078440571Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 15:37:12.078585 containerd[1992]: time="2025-02-13T15:37:12.078570175Z" level=info msg="metadata content store policy set" policy=shared Feb 13 15:37:12.083569 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 15:37:12.083811 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Feb 13 15:37:12.083969 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Feb 13 15:37:12.083969 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] Starting Core Agent Feb 13 15:37:12.083969 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [amazon-ssm-agent] registrar detected. Attempting registration Feb 13 15:37:12.083969 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [Registrar] Starting registrar module Feb 13 15:37:12.083969 amazon-ssm-agent[2027]: 2025-02-13 15:37:11 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Feb 13 15:37:12.084588 amazon-ssm-agent[2027]: 2025-02-13 15:37:12 INFO [EC2Identity] EC2 registration was successful. Feb 13 15:37:12.084588 amazon-ssm-agent[2027]: 2025-02-13 15:37:12 INFO [CredentialRefresher] credentialRefresher has started Feb 13 15:37:12.084588 amazon-ssm-agent[2027]: 2025-02-13 15:37:12 INFO [CredentialRefresher] Starting credentials refresher loop Feb 13 15:37:12.084588 amazon-ssm-agent[2027]: 2025-02-13 15:37:12 INFO EC2RoleProvider Successfully connected with instance profile role credentials Feb 13 15:37:12.086285 containerd[1992]: time="2025-02-13T15:37:12.086245112Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 15:37:12.088159 amazon-ssm-agent[2027]: 2025-02-13 15:37:12 INFO [CredentialRefresher] Next credential rotation will be in 31.791647699966667 minutes Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.086710865Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.086754086Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.086778020Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.086798304Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.086970817Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087407619Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087522974Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087544654Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087565924Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087587164Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087605705Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087627664Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087647118Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088238 containerd[1992]: time="2025-02-13T15:37:12.087668641Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087688656Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087707594Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087725419Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087752095Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087771678Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087789953Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087808651Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087825572Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087846682Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087864715Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087886308Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087925694Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087950178Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.088822 containerd[1992]: time="2025-02-13T15:37:12.087967268Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.087984254Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.088001755Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.088022667Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.088052832Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.088071151Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.089349 containerd[1992]: time="2025-02-13T15:37:12.088087360Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091380103Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091427955Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091445087Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091462604Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091478882Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091497908Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091512087Z" level=info msg="NRI interface is disabled by configuration." Feb 13 15:37:12.095003 containerd[1992]: time="2025-02-13T15:37:12.091526137Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 15:37:12.095676 containerd[1992]: time="2025-02-13T15:37:12.092141275Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 15:37:12.095676 containerd[1992]: time="2025-02-13T15:37:12.092214288Z" level=info msg="Connect containerd service" Feb 13 15:37:12.095676 containerd[1992]: time="2025-02-13T15:37:12.092278777Z" level=info msg="using legacy CRI server" Feb 13 15:37:12.095676 containerd[1992]: time="2025-02-13T15:37:12.092288554Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 15:37:12.095676 containerd[1992]: time="2025-02-13T15:37:12.092442670Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 15:37:12.099134 containerd[1992]: time="2025-02-13T15:37:12.098845171Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 15:37:12.099575 containerd[1992]: time="2025-02-13T15:37:12.099523230Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 15:37:12.099717 containerd[1992]: time="2025-02-13T15:37:12.099702211Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 15:37:12.099856 containerd[1992]: time="2025-02-13T15:37:12.099823572Z" level=info msg="Start subscribing containerd event" Feb 13 15:37:12.099946 containerd[1992]: time="2025-02-13T15:37:12.099934628Z" level=info msg="Start recovering state" Feb 13 15:37:12.100100 containerd[1992]: time="2025-02-13T15:37:12.100088195Z" level=info msg="Start event monitor" Feb 13 15:37:12.100187 containerd[1992]: time="2025-02-13T15:37:12.100175951Z" level=info msg="Start snapshots syncer" Feb 13 15:37:12.100268 containerd[1992]: time="2025-02-13T15:37:12.100257477Z" level=info msg="Start cni network conf syncer for default" Feb 13 15:37:12.101804 containerd[1992]: time="2025-02-13T15:37:12.101370462Z" level=info msg="Start streaming server" Feb 13 15:37:12.101804 containerd[1992]: time="2025-02-13T15:37:12.101490687Z" level=info msg="containerd successfully booted in 0.127999s" Feb 13 15:37:12.102423 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 15:37:12.245684 sshd_keygen[1996]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 15:37:12.280573 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 15:37:12.291742 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 15:37:12.300500 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 15:37:12.300740 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 15:37:12.317045 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 15:37:12.334296 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 15:37:12.351257 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 15:37:12.356209 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 15:37:12.358172 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 15:37:13.114870 amazon-ssm-agent[2027]: 2025-02-13 15:37:13 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Feb 13 15:37:13.215612 amazon-ssm-agent[2027]: 2025-02-13 15:37:13 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2224) started Feb 13 15:37:13.315924 amazon-ssm-agent[2027]: 2025-02-13 15:37:13 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Feb 13 15:37:13.659362 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 15:37:13.663689 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 15:37:13.666230 systemd[1]: Startup finished in 9.689s (kernel) + 10.030s (userspace) = 19.719s. Feb 13 15:37:13.672896 (kubelet)[2242]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 15:37:15.140873 kubelet[2242]: E0213 15:37:15.140753 2242 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 15:37:15.146459 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 15:37:15.146775 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 15:37:18.114228 systemd-resolved[1887]: Clock change detected. Flushing caches. Feb 13 15:37:18.659297 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 15:37:18.671901 systemd[1]: Started sshd@0-172.31.20.193:22-139.178.89.65:57256.service - OpenSSH per-connection server daemon (139.178.89.65:57256). Feb 13 15:37:18.920821 sshd[2255]: Accepted publickey for core from 139.178.89.65 port 57256 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:18.926361 sshd-session[2255]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:18.939520 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 15:37:18.949730 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 15:37:18.957651 systemd-logind[1965]: New session 1 of user core. Feb 13 15:37:18.972240 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 15:37:18.981486 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 15:37:18.995708 (systemd)[2261]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 15:37:19.139132 systemd[2261]: Queued start job for default target default.target. Feb 13 15:37:19.139708 systemd[2261]: Created slice app.slice - User Application Slice. Feb 13 15:37:19.139742 systemd[2261]: Reached target paths.target - Paths. Feb 13 15:37:19.139762 systemd[2261]: Reached target timers.target - Timers. Feb 13 15:37:19.145253 systemd[2261]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 15:37:19.154747 systemd[2261]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 15:37:19.154840 systemd[2261]: Reached target sockets.target - Sockets. Feb 13 15:37:19.154859 systemd[2261]: Reached target basic.target - Basic System. Feb 13 15:37:19.154989 systemd[2261]: Reached target default.target - Main User Target. Feb 13 15:37:19.155032 systemd[2261]: Startup finished in 151ms. Feb 13 15:37:19.156251 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 15:37:19.167169 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 15:37:19.312403 systemd[1]: Started sshd@1-172.31.20.193:22-139.178.89.65:57268.service - OpenSSH per-connection server daemon (139.178.89.65:57268). Feb 13 15:37:19.487610 sshd[2273]: Accepted publickey for core from 139.178.89.65 port 57268 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:19.493417 sshd-session[2273]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:19.501335 systemd-logind[1965]: New session 2 of user core. Feb 13 15:37:19.511925 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 15:37:19.640018 sshd[2276]: Connection closed by 139.178.89.65 port 57268 Feb 13 15:37:19.640759 sshd-session[2273]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:19.647598 systemd[1]: sshd@1-172.31.20.193:22-139.178.89.65:57268.service: Deactivated successfully. Feb 13 15:37:19.654868 systemd-logind[1965]: Session 2 logged out. Waiting for processes to exit. Feb 13 15:37:19.655697 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 15:37:19.679858 systemd[1]: Started sshd@2-172.31.20.193:22-139.178.89.65:57282.service - OpenSSH per-connection server daemon (139.178.89.65:57282). Feb 13 15:37:19.682425 systemd-logind[1965]: Removed session 2. Feb 13 15:37:19.866168 sshd[2281]: Accepted publickey for core from 139.178.89.65 port 57282 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:19.867729 sshd-session[2281]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:19.873130 systemd-logind[1965]: New session 3 of user core. Feb 13 15:37:19.886431 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 15:37:20.004171 sshd[2284]: Connection closed by 139.178.89.65 port 57282 Feb 13 15:37:20.004844 sshd-session[2281]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:20.015111 systemd[1]: sshd@2-172.31.20.193:22-139.178.89.65:57282.service: Deactivated successfully. Feb 13 15:37:20.019696 systemd-logind[1965]: Session 3 logged out. Waiting for processes to exit. Feb 13 15:37:20.020764 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 15:37:20.022709 systemd-logind[1965]: Removed session 3. Feb 13 15:37:20.035928 systemd[1]: Started sshd@3-172.31.20.193:22-139.178.89.65:57288.service - OpenSSH per-connection server daemon (139.178.89.65:57288). Feb 13 15:37:20.260818 sshd[2289]: Accepted publickey for core from 139.178.89.65 port 57288 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:20.262938 sshd-session[2289]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:20.284753 systemd-logind[1965]: New session 4 of user core. Feb 13 15:37:20.306266 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 15:37:20.433929 sshd[2292]: Connection closed by 139.178.89.65 port 57288 Feb 13 15:37:20.434841 sshd-session[2289]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:20.441722 systemd[1]: sshd@3-172.31.20.193:22-139.178.89.65:57288.service: Deactivated successfully. Feb 13 15:37:20.443483 systemd-logind[1965]: Session 4 logged out. Waiting for processes to exit. Feb 13 15:37:20.446161 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 15:37:20.447430 systemd-logind[1965]: Removed session 4. Feb 13 15:37:20.463620 systemd[1]: Started sshd@4-172.31.20.193:22-139.178.89.65:57300.service - OpenSSH per-connection server daemon (139.178.89.65:57300). Feb 13 15:37:20.627646 sshd[2297]: Accepted publickey for core from 139.178.89.65 port 57300 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:20.629227 sshd-session[2297]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:20.640844 systemd-logind[1965]: New session 5 of user core. Feb 13 15:37:20.651387 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 15:37:20.775488 sudo[2301]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 15:37:20.775886 sudo[2301]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 15:37:20.795562 sudo[2301]: pam_unix(sudo:session): session closed for user root Feb 13 15:37:20.818246 sshd[2300]: Connection closed by 139.178.89.65 port 57300 Feb 13 15:37:20.819278 sshd-session[2297]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:20.832985 systemd[1]: sshd@4-172.31.20.193:22-139.178.89.65:57300.service: Deactivated successfully. Feb 13 15:37:20.843506 systemd-logind[1965]: Session 5 logged out. Waiting for processes to exit. Feb 13 15:37:20.849197 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 15:37:20.865465 systemd[1]: Started sshd@5-172.31.20.193:22-139.178.89.65:57306.service - OpenSSH per-connection server daemon (139.178.89.65:57306). Feb 13 15:37:20.867302 systemd-logind[1965]: Removed session 5. Feb 13 15:37:21.037983 sshd[2306]: Accepted publickey for core from 139.178.89.65 port 57306 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:21.040203 sshd-session[2306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:21.046135 systemd-logind[1965]: New session 6 of user core. Feb 13 15:37:21.052555 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 15:37:21.169098 sudo[2311]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 15:37:21.169618 sudo[2311]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 15:37:21.176546 sudo[2311]: pam_unix(sudo:session): session closed for user root Feb 13 15:37:21.186692 sudo[2310]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Feb 13 15:37:21.187144 sudo[2310]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 15:37:21.215585 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 15:37:21.267956 augenrules[2333]: No rules Feb 13 15:37:21.270093 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 15:37:21.270539 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 15:37:21.275020 sudo[2310]: pam_unix(sudo:session): session closed for user root Feb 13 15:37:21.297991 sshd[2309]: Connection closed by 139.178.89.65 port 57306 Feb 13 15:37:21.299483 sshd-session[2306]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:21.305920 systemd[1]: sshd@5-172.31.20.193:22-139.178.89.65:57306.service: Deactivated successfully. Feb 13 15:37:21.310768 systemd-logind[1965]: Session 6 logged out. Waiting for processes to exit. Feb 13 15:37:21.311574 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 15:37:21.314208 systemd-logind[1965]: Removed session 6. Feb 13 15:37:21.333677 systemd[1]: Started sshd@6-172.31.20.193:22-139.178.89.65:57310.service - OpenSSH per-connection server daemon (139.178.89.65:57310). Feb 13 15:37:21.514207 sshd[2342]: Accepted publickey for core from 139.178.89.65 port 57310 ssh2: RSA SHA256:v7hTrtZ9/NhiAvXSp1iZfOxZYI4fXxME+gLHhLHyxgM Feb 13 15:37:21.515848 sshd-session[2342]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 15:37:21.521883 systemd-logind[1965]: New session 7 of user core. Feb 13 15:37:21.529539 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 15:37:21.630697 sudo[2346]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 15:37:21.631108 sudo[2346]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 15:37:23.117739 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 15:37:23.131431 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 15:37:23.176966 systemd[1]: Reloading requested from client PID 2387 ('systemctl') (unit session-7.scope)... Feb 13 15:37:23.177034 systemd[1]: Reloading... Feb 13 15:37:23.360114 zram_generator::config[2428]: No configuration found. Feb 13 15:37:23.553470 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 15:37:23.658146 systemd[1]: Reloading finished in 480 ms. Feb 13 15:37:23.734142 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Feb 13 15:37:23.734497 systemd[1]: kubelet.service: Failed with result 'signal'. Feb 13 15:37:23.735277 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 15:37:23.746492 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 15:37:23.937274 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 15:37:23.950741 (kubelet)[2499]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 15:37:24.013827 kubelet[2499]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 15:37:24.015130 kubelet[2499]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 13 15:37:24.015130 kubelet[2499]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 15:37:24.015130 kubelet[2499]: I0213 15:37:24.014325 2499 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 15:37:24.217250 kubelet[2499]: I0213 15:37:24.217132 2499 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Feb 13 15:37:24.217250 kubelet[2499]: I0213 15:37:24.217167 2499 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 15:37:24.218049 kubelet[2499]: I0213 15:37:24.217515 2499 server.go:919] "Client rotation is on, will bootstrap in background" Feb 13 15:37:24.258783 kubelet[2499]: I0213 15:37:24.258733 2499 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 15:37:24.278740 kubelet[2499]: I0213 15:37:24.278698 2499 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 15:37:24.279854 kubelet[2499]: I0213 15:37:24.279824 2499 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 15:37:24.280257 kubelet[2499]: I0213 15:37:24.280172 2499 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Feb 13 15:37:24.281471 kubelet[2499]: I0213 15:37:24.281442 2499 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 15:37:24.281471 kubelet[2499]: I0213 15:37:24.281473 2499 container_manager_linux.go:301] "Creating device plugin manager" Feb 13 15:37:24.281710 kubelet[2499]: I0213 15:37:24.281613 2499 state_mem.go:36] "Initialized new in-memory state store" Feb 13 15:37:24.281963 kubelet[2499]: I0213 15:37:24.281814 2499 kubelet.go:396] "Attempting to sync node with API server" Feb 13 15:37:24.281963 kubelet[2499]: I0213 15:37:24.281960 2499 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 15:37:24.282102 kubelet[2499]: I0213 15:37:24.282001 2499 kubelet.go:312] "Adding apiserver pod source" Feb 13 15:37:24.282102 kubelet[2499]: I0213 15:37:24.282021 2499 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 15:37:24.293164 kubelet[2499]: E0213 15:37:24.292344 2499 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:24.293164 kubelet[2499]: E0213 15:37:24.292807 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:24.294079 kubelet[2499]: I0213 15:37:24.293966 2499 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 15:37:24.299830 kubelet[2499]: I0213 15:37:24.299778 2499 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 15:37:24.300062 kubelet[2499]: W0213 15:37:24.300007 2499 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 15:37:24.301615 kubelet[2499]: I0213 15:37:24.300824 2499 server.go:1256] "Started kubelet" Feb 13 15:37:24.303544 kubelet[2499]: I0213 15:37:24.303201 2499 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 15:37:24.311307 kubelet[2499]: I0213 15:37:24.311278 2499 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 15:37:24.315835 kubelet[2499]: I0213 15:37:24.315807 2499 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 15:37:24.316787 kubelet[2499]: I0213 15:37:24.316768 2499 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 15:37:24.332347 kubelet[2499]: I0213 15:37:24.332310 2499 volume_manager.go:291] "Starting Kubelet Volume Manager" Feb 13 15:37:24.335018 kubelet[2499]: I0213 15:37:24.334499 2499 server.go:461] "Adding debug handlers to kubelet server" Feb 13 15:37:24.341529 kubelet[2499]: E0213 15:37:24.337820 2499 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.20.193.1823ce9a8e2dbf27 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.20.193,UID:172.31.20.193,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172.31.20.193,},FirstTimestamp:2025-02-13 15:37:24.300791591 +0000 UTC m=+0.344437091,LastTimestamp:2025-02-13 15:37:24.300791591 +0000 UTC m=+0.344437091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.20.193,}" Feb 13 15:37:24.343770 kubelet[2499]: I0213 15:37:24.343739 2499 factory.go:221] Registration of the systemd container factory successfully Feb 13 15:37:24.344014 kubelet[2499]: I0213 15:37:24.343908 2499 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 15:37:24.348095 kubelet[2499]: I0213 15:37:24.344816 2499 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Feb 13 15:37:24.348095 kubelet[2499]: I0213 15:37:24.345011 2499 reconciler_new.go:29] "Reconciler: start to sync state" Feb 13 15:37:24.348714 kubelet[2499]: I0213 15:37:24.348694 2499 factory.go:221] Registration of the containerd container factory successfully Feb 13 15:37:24.366688 kubelet[2499]: E0213 15:37:24.366655 2499 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 15:37:24.391384 kubelet[2499]: W0213 15:37:24.391353 2499 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: nodes "172.31.20.193" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 13 15:37:24.391384 kubelet[2499]: E0213 15:37:24.391391 2499 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: nodes "172.31.20.193" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 13 15:37:24.391571 kubelet[2499]: E0213 15:37:24.391444 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"172.31.20.193\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Feb 13 15:37:24.391571 kubelet[2499]: W0213 15:37:24.391483 2499 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 13 15:37:24.391571 kubelet[2499]: E0213 15:37:24.391496 2499 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 13 15:37:24.391696 kubelet[2499]: W0213 15:37:24.391629 2499 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 13 15:37:24.391696 kubelet[2499]: E0213 15:37:24.391645 2499 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 13 15:37:24.396362 kubelet[2499]: E0213 15:37:24.396326 2499 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.20.193.1823ce9a921a5afd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.20.193,UID:172.31.20.193,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:172.31.20.193,},FirstTimestamp:2025-02-13 15:37:24.366629629 +0000 UTC m=+0.410275135,LastTimestamp:2025-02-13 15:37:24.366629629 +0000 UTC m=+0.410275135,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.20.193,}" Feb 13 15:37:24.398825 kubelet[2499]: I0213 15:37:24.398803 2499 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 13 15:37:24.398950 kubelet[2499]: I0213 15:37:24.398941 2499 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 13 15:37:24.400854 kubelet[2499]: I0213 15:37:24.400505 2499 state_mem.go:36] "Initialized new in-memory state store" Feb 13 15:37:24.405421 kubelet[2499]: I0213 15:37:24.405087 2499 policy_none.go:49] "None policy: Start" Feb 13 15:37:24.407825 kubelet[2499]: I0213 15:37:24.407803 2499 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 13 15:37:24.408634 kubelet[2499]: I0213 15:37:24.408025 2499 state_mem.go:35] "Initializing new in-memory state store" Feb 13 15:37:24.425609 kubelet[2499]: I0213 15:37:24.425564 2499 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 15:37:24.425884 kubelet[2499]: I0213 15:37:24.425856 2499 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 15:37:24.434106 kubelet[2499]: E0213 15:37:24.432303 2499 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.20.193.1823ce9a93f5f5af default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.20.193,UID:172.31.20.193,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node 172.31.20.193 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:172.31.20.193,},FirstTimestamp:2025-02-13 15:37:24.397798831 +0000 UTC m=+0.441444318,LastTimestamp:2025-02-13 15:37:24.397798831 +0000 UTC m=+0.441444318,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.20.193,}" Feb 13 15:37:24.436419 kubelet[2499]: I0213 15:37:24.436394 2499 kubelet_node_status.go:73] "Attempting to register node" node="172.31.20.193" Feb 13 15:37:24.438262 kubelet[2499]: E0213 15:37:24.438224 2499 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172.31.20.193\" not found" Feb 13 15:37:24.482744 kubelet[2499]: I0213 15:37:24.482575 2499 kubelet_node_status.go:76] "Successfully registered node" node="172.31.20.193" Feb 13 15:37:24.491936 kubelet[2499]: I0213 15:37:24.491898 2499 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 15:37:24.493871 kubelet[2499]: I0213 15:37:24.493842 2499 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 15:37:24.493981 kubelet[2499]: I0213 15:37:24.493882 2499 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 13 15:37:24.493981 kubelet[2499]: I0213 15:37:24.493903 2499 kubelet.go:2329] "Starting kubelet main sync loop" Feb 13 15:37:24.494210 kubelet[2499]: E0213 15:37:24.494013 2499 kubelet.go:2353] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Feb 13 15:37:24.509261 kubelet[2499]: E0213 15:37:24.509223 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:24.609843 kubelet[2499]: E0213 15:37:24.609802 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:24.710037 kubelet[2499]: E0213 15:37:24.709998 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:24.810855 kubelet[2499]: E0213 15:37:24.810673 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:24.911350 kubelet[2499]: E0213 15:37:24.911300 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.012313 kubelet[2499]: E0213 15:37:25.012037 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.112920 kubelet[2499]: E0213 15:37:25.112805 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.213577 kubelet[2499]: E0213 15:37:25.213527 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.221868 kubelet[2499]: I0213 15:37:25.221761 2499 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 13 15:37:25.222039 kubelet[2499]: W0213 15:37:25.222008 2499 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.RuntimeClass ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Feb 13 15:37:25.293599 kubelet[2499]: E0213 15:37:25.293545 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:25.313763 kubelet[2499]: E0213 15:37:25.313708 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.414600 kubelet[2499]: E0213 15:37:25.414549 2499 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.20.193\" not found" Feb 13 15:37:25.486868 sudo[2346]: pam_unix(sudo:session): session closed for user root Feb 13 15:37:25.509723 sshd[2345]: Connection closed by 139.178.89.65 port 57310 Feb 13 15:37:25.511435 sshd-session[2342]: pam_unix(sshd:session): session closed for user core Feb 13 15:37:25.515866 systemd[1]: sshd@6-172.31.20.193:22-139.178.89.65:57310.service: Deactivated successfully. Feb 13 15:37:25.522223 kubelet[2499]: I0213 15:37:25.518627 2499 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Feb 13 15:37:25.522223 kubelet[2499]: I0213 15:37:25.520381 2499 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Feb 13 15:37:25.522360 containerd[1992]: time="2025-02-13T15:37:25.519269262Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 15:37:25.523271 systemd-logind[1965]: Session 7 logged out. Waiting for processes to exit. Feb 13 15:37:25.523736 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 15:37:25.526670 systemd-logind[1965]: Removed session 7. Feb 13 15:37:26.293821 kubelet[2499]: E0213 15:37:26.293756 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:26.293821 kubelet[2499]: I0213 15:37:26.293761 2499 apiserver.go:52] "Watching apiserver" Feb 13 15:37:26.304136 kubelet[2499]: I0213 15:37:26.304096 2499 topology_manager.go:215] "Topology Admit Handler" podUID="0da55bd1-e851-46c5-9d38-044ed9a3c4ca" podNamespace="calico-system" podName="calico-node-4hcb2" Feb 13 15:37:26.305151 kubelet[2499]: I0213 15:37:26.304274 2499 topology_manager.go:215] "Topology Admit Handler" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" podNamespace="calico-system" podName="csi-node-driver-kxrfd" Feb 13 15:37:26.305151 kubelet[2499]: I0213 15:37:26.304339 2499 topology_manager.go:215] "Topology Admit Handler" podUID="3f39489a-d9fe-4213-ba07-517dcacd55a6" podNamespace="kube-system" podName="kube-proxy-vpp59" Feb 13 15:37:26.305339 kubelet[2499]: E0213 15:37:26.305305 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:26.346551 kubelet[2499]: I0213 15:37:26.346482 2499 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Feb 13 15:37:26.359630 kubelet[2499]: I0213 15:37:26.359583 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-node-certs\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.359630 kubelet[2499]: I0213 15:37:26.359636 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-cni-net-dir\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.359800 kubelet[2499]: I0213 15:37:26.359675 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9afcbeb8-f8e4-4f2f-9b35-6cef169285f0-registration-dir\") pod \"csi-node-driver-kxrfd\" (UID: \"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0\") " pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:26.359800 kubelet[2499]: I0213 15:37:26.359701 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3f39489a-d9fe-4213-ba07-517dcacd55a6-lib-modules\") pod \"kube-proxy-vpp59\" (UID: \"3f39489a-d9fe-4213-ba07-517dcacd55a6\") " pod="kube-system/kube-proxy-vpp59" Feb 13 15:37:26.359800 kubelet[2499]: I0213 15:37:26.359728 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9afcbeb8-f8e4-4f2f-9b35-6cef169285f0-kubelet-dir\") pod \"csi-node-driver-kxrfd\" (UID: \"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0\") " pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:26.359800 kubelet[2499]: I0213 15:37:26.359755 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-lib-modules\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.359800 kubelet[2499]: I0213 15:37:26.359786 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-var-run-calico\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360102 kubelet[2499]: I0213 15:37:26.359813 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-var-lib-calico\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360102 kubelet[2499]: I0213 15:37:26.359848 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-flexvol-driver-host\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360102 kubelet[2499]: I0213 15:37:26.359879 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fbgb\" (UniqueName: \"kubernetes.io/projected/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-kube-api-access-6fbgb\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360102 kubelet[2499]: I0213 15:37:26.359909 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxb8w\" (UniqueName: \"kubernetes.io/projected/3f39489a-d9fe-4213-ba07-517dcacd55a6-kube-api-access-hxb8w\") pod \"kube-proxy-vpp59\" (UID: \"3f39489a-d9fe-4213-ba07-517dcacd55a6\") " pod="kube-system/kube-proxy-vpp59" Feb 13 15:37:26.360102 kubelet[2499]: I0213 15:37:26.359938 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-policysync\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360330 kubelet[2499]: I0213 15:37:26.359966 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-tigera-ca-bundle\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360330 kubelet[2499]: I0213 15:37:26.359995 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-cni-bin-dir\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360330 kubelet[2499]: I0213 15:37:26.360027 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-cni-log-dir\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360330 kubelet[2499]: I0213 15:37:26.360056 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3f39489a-d9fe-4213-ba07-517dcacd55a6-kube-proxy\") pod \"kube-proxy-vpp59\" (UID: \"3f39489a-d9fe-4213-ba07-517dcacd55a6\") " pod="kube-system/kube-proxy-vpp59" Feb 13 15:37:26.360330 kubelet[2499]: I0213 15:37:26.360102 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0da55bd1-e851-46c5-9d38-044ed9a3c4ca-xtables-lock\") pod \"calico-node-4hcb2\" (UID: \"0da55bd1-e851-46c5-9d38-044ed9a3c4ca\") " pod="calico-system/calico-node-4hcb2" Feb 13 15:37:26.360488 kubelet[2499]: I0213 15:37:26.360136 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/9afcbeb8-f8e4-4f2f-9b35-6cef169285f0-varrun\") pod \"csi-node-driver-kxrfd\" (UID: \"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0\") " pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:26.360488 kubelet[2499]: I0213 15:37:26.360169 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9afcbeb8-f8e4-4f2f-9b35-6cef169285f0-socket-dir\") pod \"csi-node-driver-kxrfd\" (UID: \"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0\") " pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:26.360488 kubelet[2499]: I0213 15:37:26.360219 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtgnp\" (UniqueName: \"kubernetes.io/projected/9afcbeb8-f8e4-4f2f-9b35-6cef169285f0-kube-api-access-vtgnp\") pod \"csi-node-driver-kxrfd\" (UID: \"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0\") " pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:26.360488 kubelet[2499]: I0213 15:37:26.360249 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3f39489a-d9fe-4213-ba07-517dcacd55a6-xtables-lock\") pod \"kube-proxy-vpp59\" (UID: \"3f39489a-d9fe-4213-ba07-517dcacd55a6\") " pod="kube-system/kube-proxy-vpp59" Feb 13 15:37:26.463606 kubelet[2499]: E0213 15:37:26.463567 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.463606 kubelet[2499]: W0213 15:37:26.463589 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.463930 kubelet[2499]: E0213 15:37:26.463624 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.463930 kubelet[2499]: E0213 15:37:26.463922 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.464066 kubelet[2499]: W0213 15:37:26.463935 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.464066 kubelet[2499]: E0213 15:37:26.463955 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.464694 kubelet[2499]: E0213 15:37:26.464313 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.464694 kubelet[2499]: W0213 15:37:26.464324 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.464694 kubelet[2499]: E0213 15:37:26.464340 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.464832 kubelet[2499]: E0213 15:37:26.464750 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.464832 kubelet[2499]: W0213 15:37:26.464760 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.464832 kubelet[2499]: E0213 15:37:26.464777 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.464977 kubelet[2499]: E0213 15:37:26.464963 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.464977 kubelet[2499]: W0213 15:37:26.464972 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.465089 kubelet[2499]: E0213 15:37:26.464987 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.465585 kubelet[2499]: E0213 15:37:26.465186 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.465585 kubelet[2499]: W0213 15:37:26.465198 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.465585 kubelet[2499]: E0213 15:37:26.465216 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.465585 kubelet[2499]: E0213 15:37:26.465403 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.465585 kubelet[2499]: W0213 15:37:26.465412 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.465585 kubelet[2499]: E0213 15:37:26.465426 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.466111 kubelet[2499]: E0213 15:37:26.465600 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.466111 kubelet[2499]: W0213 15:37:26.465609 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.466111 kubelet[2499]: E0213 15:37:26.465623 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.466111 kubelet[2499]: E0213 15:37:26.465828 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.466111 kubelet[2499]: W0213 15:37:26.465837 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.466111 kubelet[2499]: E0213 15:37:26.465853 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.466463 kubelet[2499]: E0213 15:37:26.466322 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.466463 kubelet[2499]: W0213 15:37:26.466333 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.466463 kubelet[2499]: E0213 15:37:26.466350 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.466695 kubelet[2499]: E0213 15:37:26.466548 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.466695 kubelet[2499]: W0213 15:37:26.466557 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.466695 kubelet[2499]: E0213 15:37:26.466588 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.472408 kubelet[2499]: E0213 15:37:26.466875 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.472408 kubelet[2499]: W0213 15:37:26.472215 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.472408 kubelet[2499]: E0213 15:37:26.472271 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.472593 kubelet[2499]: E0213 15:37:26.472580 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.472593 kubelet[2499]: W0213 15:37:26.472593 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.472890 kubelet[2499]: E0213 15:37:26.472633 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.472890 kubelet[2499]: E0213 15:37:26.472809 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.472890 kubelet[2499]: W0213 15:37:26.472819 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.472890 kubelet[2499]: E0213 15:37:26.472845 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.473152 kubelet[2499]: E0213 15:37:26.473015 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.473152 kubelet[2499]: W0213 15:37:26.473025 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.473152 kubelet[2499]: E0213 15:37:26.473059 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.473367 kubelet[2499]: E0213 15:37:26.473247 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.473367 kubelet[2499]: W0213 15:37:26.473257 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.473367 kubelet[2499]: E0213 15:37:26.473342 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.473580 kubelet[2499]: E0213 15:37:26.473497 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.473580 kubelet[2499]: W0213 15:37:26.473506 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.473580 kubelet[2499]: E0213 15:37:26.473525 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.473756 kubelet[2499]: E0213 15:37:26.473735 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.473756 kubelet[2499]: W0213 15:37:26.473745 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.474564 kubelet[2499]: E0213 15:37:26.473774 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.474939 kubelet[2499]: E0213 15:37:26.474919 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.474939 kubelet[2499]: W0213 15:37:26.474935 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.475231 kubelet[2499]: E0213 15:37:26.474958 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.475375 kubelet[2499]: E0213 15:37:26.475358 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.475375 kubelet[2499]: W0213 15:37:26.475373 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.475569 kubelet[2499]: E0213 15:37:26.475554 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.475767 kubelet[2499]: E0213 15:37:26.475750 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.475767 kubelet[2499]: W0213 15:37:26.475764 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.475882 kubelet[2499]: E0213 15:37:26.475867 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.476039 kubelet[2499]: E0213 15:37:26.476021 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.476039 kubelet[2499]: W0213 15:37:26.476035 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.476192 kubelet[2499]: E0213 15:37:26.476179 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.476353 kubelet[2499]: E0213 15:37:26.476337 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.476353 kubelet[2499]: W0213 15:37:26.476350 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.476554 kubelet[2499]: E0213 15:37:26.476543 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.477052 kubelet[2499]: E0213 15:37:26.477033 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.477052 kubelet[2499]: W0213 15:37:26.477048 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.477225 kubelet[2499]: E0213 15:37:26.477147 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.477434 kubelet[2499]: E0213 15:37:26.477417 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.477434 kubelet[2499]: W0213 15:37:26.477430 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.477646 kubelet[2499]: E0213 15:37:26.477518 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.477823 kubelet[2499]: E0213 15:37:26.477810 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.478007 kubelet[2499]: W0213 15:37:26.477824 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.478007 kubelet[2499]: E0213 15:37:26.477940 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.478276 kubelet[2499]: E0213 15:37:26.478205 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.478276 kubelet[2499]: W0213 15:37:26.478218 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.478276 kubelet[2499]: E0213 15:37:26.478240 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.478497 kubelet[2499]: E0213 15:37:26.478468 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.478497 kubelet[2499]: W0213 15:37:26.478480 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.478497 kubelet[2499]: E0213 15:37:26.478496 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.478758 kubelet[2499]: E0213 15:37:26.478743 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.478758 kubelet[2499]: W0213 15:37:26.478755 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.478951 kubelet[2499]: E0213 15:37:26.478772 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.486391 kubelet[2499]: E0213 15:37:26.486277 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.486391 kubelet[2499]: W0213 15:37:26.486300 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.486391 kubelet[2499]: E0213 15:37:26.486327 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.582761 kubelet[2499]: E0213 15:37:26.578755 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.582761 kubelet[2499]: W0213 15:37:26.578782 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.582761 kubelet[2499]: E0213 15:37:26.578813 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.590104 kubelet[2499]: E0213 15:37:26.589617 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.600437 kubelet[2499]: W0213 15:37:26.598107 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.600437 kubelet[2499]: E0213 15:37:26.598157 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.606273 kubelet[2499]: E0213 15:37:26.600695 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.606273 kubelet[2499]: W0213 15:37:26.600735 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.606273 kubelet[2499]: E0213 15:37:26.600762 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.608446 kubelet[2499]: E0213 15:37:26.608424 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.609092 kubelet[2499]: W0213 15:37:26.608556 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.609092 kubelet[2499]: E0213 15:37:26.608588 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.614257 kubelet[2499]: E0213 15:37:26.612149 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.614257 kubelet[2499]: W0213 15:37:26.612169 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.614257 kubelet[2499]: E0213 15:37:26.612209 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.620120 containerd[1992]: time="2025-02-13T15:37:26.615814043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vpp59,Uid:3f39489a-d9fe-4213-ba07-517dcacd55a6,Namespace:kube-system,Attempt:0,}" Feb 13 15:37:26.624185 kubelet[2499]: E0213 15:37:26.624160 2499 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 15:37:26.626144 kubelet[2499]: W0213 15:37:26.626111 2499 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 15:37:26.628090 kubelet[2499]: E0213 15:37:26.626279 2499 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 15:37:26.911407 containerd[1992]: time="2025-02-13T15:37:26.911351564Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-4hcb2,Uid:0da55bd1-e851-46c5-9d38-044ed9a3c4ca,Namespace:calico-system,Attempt:0,}" Feb 13 15:37:27.276422 containerd[1992]: time="2025-02-13T15:37:27.276107814Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 15:37:27.279459 containerd[1992]: time="2025-02-13T15:37:27.279409183Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Feb 13 15:37:27.280709 containerd[1992]: time="2025-02-13T15:37:27.280670501Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 15:37:27.282992 containerd[1992]: time="2025-02-13T15:37:27.282951825Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 15:37:27.284124 containerd[1992]: time="2025-02-13T15:37:27.284025672Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 15:37:27.287897 containerd[1992]: time="2025-02-13T15:37:27.287842839Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 15:37:27.291097 containerd[1992]: time="2025-02-13T15:37:27.288908738Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 377.43839ms" Feb 13 15:37:27.291097 containerd[1992]: time="2025-02-13T15:37:27.291005791Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 674.983259ms" Feb 13 15:37:27.294400 kubelet[2499]: E0213 15:37:27.294299 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:27.506485 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3663233420.mount: Deactivated successfully. Feb 13 15:37:27.536727 containerd[1992]: time="2025-02-13T15:37:27.536011076Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:37:27.536979 containerd[1992]: time="2025-02-13T15:37:27.536884534Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:37:27.537156 containerd[1992]: time="2025-02-13T15:37:27.537107306Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:27.539310 containerd[1992]: time="2025-02-13T15:37:27.539136595Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:27.540237 containerd[1992]: time="2025-02-13T15:37:27.532629493Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:37:27.540344 containerd[1992]: time="2025-02-13T15:37:27.540239695Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:37:27.540344 containerd[1992]: time="2025-02-13T15:37:27.540270368Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:27.540464 containerd[1992]: time="2025-02-13T15:37:27.540387137Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:27.718714 containerd[1992]: time="2025-02-13T15:37:27.718675825Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-4hcb2,Uid:0da55bd1-e851-46c5-9d38-044ed9a3c4ca,Namespace:calico-system,Attempt:0,} returns sandbox id \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\"" Feb 13 15:37:27.723520 containerd[1992]: time="2025-02-13T15:37:27.722289285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vpp59,Uid:3f39489a-d9fe-4213-ba07-517dcacd55a6,Namespace:kube-system,Attempt:0,} returns sandbox id \"37b61f09da3a3f23b512dd6dc3db6dbc1cebb3579d359f0019a3f4ddc4adb5c7\"" Feb 13 15:37:27.729337 containerd[1992]: time="2025-02-13T15:37:27.729285212Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Feb 13 15:37:28.296092 kubelet[2499]: E0213 15:37:28.295227 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:28.495988 kubelet[2499]: E0213 15:37:28.495951 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:29.296916 kubelet[2499]: E0213 15:37:29.296447 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:29.303861 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1789108274.mount: Deactivated successfully. Feb 13 15:37:29.815605 containerd[1992]: time="2025-02-13T15:37:29.815552942Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:29.818263 containerd[1992]: time="2025-02-13T15:37:29.818202040Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=6855343" Feb 13 15:37:29.819418 containerd[1992]: time="2025-02-13T15:37:29.819359450Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:29.823641 containerd[1992]: time="2025-02-13T15:37:29.823572246Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:29.825837 containerd[1992]: time="2025-02-13T15:37:29.824550527Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 2.095205876s" Feb 13 15:37:29.825837 containerd[1992]: time="2025-02-13T15:37:29.824590768Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Feb 13 15:37:29.826766 containerd[1992]: time="2025-02-13T15:37:29.826741456Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.14\"" Feb 13 15:37:29.828371 containerd[1992]: time="2025-02-13T15:37:29.828334361Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Feb 13 15:37:29.856617 containerd[1992]: time="2025-02-13T15:37:29.856569573Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674\"" Feb 13 15:37:29.859869 containerd[1992]: time="2025-02-13T15:37:29.859827629Z" level=info msg="StartContainer for \"722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674\"" Feb 13 15:37:29.956556 containerd[1992]: time="2025-02-13T15:37:29.956499894Z" level=info msg="StartContainer for \"722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674\" returns successfully" Feb 13 15:37:30.072629 containerd[1992]: time="2025-02-13T15:37:30.072470260Z" level=info msg="shim disconnected" id=722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674 namespace=k8s.io Feb 13 15:37:30.072629 containerd[1992]: time="2025-02-13T15:37:30.072531198Z" level=warning msg="cleaning up after shim disconnected" id=722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674 namespace=k8s.io Feb 13 15:37:30.072629 containerd[1992]: time="2025-02-13T15:37:30.072546817Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 15:37:30.243233 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-722f7d5487ee8188e074a512d87c4ae433c774ac80e82b49ccb702c1d1f9d674-rootfs.mount: Deactivated successfully. Feb 13 15:37:30.297118 kubelet[2499]: E0213 15:37:30.297044 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:30.495923 kubelet[2499]: E0213 15:37:30.495020 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:31.065833 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2921562179.mount: Deactivated successfully. Feb 13 15:37:31.298643 kubelet[2499]: E0213 15:37:31.298303 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:31.813702 containerd[1992]: time="2025-02-13T15:37:31.813639398Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.14\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:31.815294 containerd[1992]: time="2025-02-13T15:37:31.815060843Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.14: active requests=0, bytes read=28620592" Feb 13 15:37:31.817114 containerd[1992]: time="2025-02-13T15:37:31.816732232Z" level=info msg="ImageCreate event name:\"sha256:609f2866f1e52a5f0d2651e1206db6aeb38e8c3f91175abcfaf7e87381e5cce2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:31.821130 containerd[1992]: time="2025-02-13T15:37:31.821089614Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:197988595a902751e4e570a5e4d74182f12d83c1d175c1e79aa020f358f6535b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:31.821995 containerd[1992]: time="2025-02-13T15:37:31.821961700Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.14\" with image id \"sha256:609f2866f1e52a5f0d2651e1206db6aeb38e8c3f91175abcfaf7e87381e5cce2\", repo tag \"registry.k8s.io/kube-proxy:v1.29.14\", repo digest \"registry.k8s.io/kube-proxy@sha256:197988595a902751e4e570a5e4d74182f12d83c1d175c1e79aa020f358f6535b\", size \"28619611\" in 1.994962457s" Feb 13 15:37:31.822165 containerd[1992]: time="2025-02-13T15:37:31.822145221Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.14\" returns image reference \"sha256:609f2866f1e52a5f0d2651e1206db6aeb38e8c3f91175abcfaf7e87381e5cce2\"" Feb 13 15:37:31.824104 containerd[1992]: time="2025-02-13T15:37:31.824005357Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Feb 13 15:37:31.828091 containerd[1992]: time="2025-02-13T15:37:31.825829537Z" level=info msg="CreateContainer within sandbox \"37b61f09da3a3f23b512dd6dc3db6dbc1cebb3579d359f0019a3f4ddc4adb5c7\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 15:37:31.853053 containerd[1992]: time="2025-02-13T15:37:31.852977258Z" level=info msg="CreateContainer within sandbox \"37b61f09da3a3f23b512dd6dc3db6dbc1cebb3579d359f0019a3f4ddc4adb5c7\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"2d84315cc7948630bea2f9b0803b8243484ee9b8d45fbd2e45c709028b91096a\"" Feb 13 15:37:31.854409 containerd[1992]: time="2025-02-13T15:37:31.854351279Z" level=info msg="StartContainer for \"2d84315cc7948630bea2f9b0803b8243484ee9b8d45fbd2e45c709028b91096a\"" Feb 13 15:37:31.939408 containerd[1992]: time="2025-02-13T15:37:31.939356357Z" level=info msg="StartContainer for \"2d84315cc7948630bea2f9b0803b8243484ee9b8d45fbd2e45c709028b91096a\" returns successfully" Feb 13 15:37:32.299293 kubelet[2499]: E0213 15:37:32.299236 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:32.501254 kubelet[2499]: E0213 15:37:32.500632 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:33.303697 kubelet[2499]: E0213 15:37:33.303536 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:34.304897 kubelet[2499]: E0213 15:37:34.304855 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:34.495142 kubelet[2499]: E0213 15:37:34.494649 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:35.305356 kubelet[2499]: E0213 15:37:35.305222 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:36.274939 containerd[1992]: time="2025-02-13T15:37:36.274891028Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:36.276432 containerd[1992]: time="2025-02-13T15:37:36.276294224Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Feb 13 15:37:36.277894 containerd[1992]: time="2025-02-13T15:37:36.277611823Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:36.280981 containerd[1992]: time="2025-02-13T15:37:36.280722053Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:36.282099 containerd[1992]: time="2025-02-13T15:37:36.281867889Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 4.457826879s" Feb 13 15:37:36.282099 containerd[1992]: time="2025-02-13T15:37:36.281904336Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Feb 13 15:37:36.284201 containerd[1992]: time="2025-02-13T15:37:36.284120457Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 15:37:36.305958 kubelet[2499]: E0213 15:37:36.305892 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:36.306978 containerd[1992]: time="2025-02-13T15:37:36.306014459Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404\"" Feb 13 15:37:36.306978 containerd[1992]: time="2025-02-13T15:37:36.306680120Z" level=info msg="StartContainer for \"a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404\"" Feb 13 15:37:36.387271 containerd[1992]: time="2025-02-13T15:37:36.387097529Z" level=info msg="StartContainer for \"a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404\" returns successfully" Feb 13 15:37:36.496666 kubelet[2499]: E0213 15:37:36.494660 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:36.604552 kubelet[2499]: I0213 15:37:36.604430 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-vpp59" podStartSLOduration=8.508488595 podStartE2EDuration="12.604368265s" podCreationTimestamp="2025-02-13 15:37:24 +0000 UTC" firstStartedPulling="2025-02-13 15:37:27.727150961 +0000 UTC m=+3.770796455" lastFinishedPulling="2025-02-13 15:37:31.823030629 +0000 UTC m=+7.866676125" observedRunningTime="2025-02-13 15:37:32.605997775 +0000 UTC m=+8.649643281" watchObservedRunningTime="2025-02-13 15:37:36.604368265 +0000 UTC m=+12.648013770" Feb 13 15:37:37.307032 kubelet[2499]: E0213 15:37:37.306967 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:37.433535 containerd[1992]: time="2025-02-13T15:37:37.433312201Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 15:37:37.465289 kubelet[2499]: I0213 15:37:37.464851 2499 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Feb 13 15:37:37.500022 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404-rootfs.mount: Deactivated successfully. Feb 13 15:37:37.747901 containerd[1992]: time="2025-02-13T15:37:37.747828936Z" level=info msg="shim disconnected" id=a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404 namespace=k8s.io Feb 13 15:37:37.747901 containerd[1992]: time="2025-02-13T15:37:37.747895604Z" level=warning msg="cleaning up after shim disconnected" id=a0b1d250e35d12e999edc1a6fde229577e9f69839b41111039b82846205d9404 namespace=k8s.io Feb 13 15:37:37.747901 containerd[1992]: time="2025-02-13T15:37:37.747907132Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 15:37:38.307331 kubelet[2499]: E0213 15:37:38.307278 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:38.499116 containerd[1992]: time="2025-02-13T15:37:38.498636966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:0,}" Feb 13 15:37:38.578841 containerd[1992]: time="2025-02-13T15:37:38.578606516Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Feb 13 15:37:38.632348 containerd[1992]: time="2025-02-13T15:37:38.632290946Z" level=error msg="Failed to destroy network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:38.635492 containerd[1992]: time="2025-02-13T15:37:38.635436589Z" level=error msg="encountered an error cleaning up failed sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:38.635617 containerd[1992]: time="2025-02-13T15:37:38.635535300Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:38.635998 kubelet[2499]: E0213 15:37:38.635905 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:38.636092 kubelet[2499]: E0213 15:37:38.636065 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:38.636149 kubelet[2499]: E0213 15:37:38.636106 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:38.636372 kubelet[2499]: E0213 15:37:38.636352 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:38.636757 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe-shm.mount: Deactivated successfully. Feb 13 15:37:39.307452 kubelet[2499]: E0213 15:37:39.307392 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:39.574552 kubelet[2499]: I0213 15:37:39.574387 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe" Feb 13 15:37:39.575284 containerd[1992]: time="2025-02-13T15:37:39.575240462Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:39.578286 containerd[1992]: time="2025-02-13T15:37:39.575631438Z" level=info msg="Ensure that sandbox af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe in task-service has been cleanup successfully" Feb 13 15:37:39.578286 containerd[1992]: time="2025-02-13T15:37:39.575813643Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:39.578286 containerd[1992]: time="2025-02-13T15:37:39.575830459Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:39.578286 containerd[1992]: time="2025-02-13T15:37:39.576447757Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:1,}" Feb 13 15:37:39.578421 systemd[1]: run-netns-cni\x2d970a7fa9\x2dee6f\x2dde00\x2d4e6d\x2d1f2abe7dcc9c.mount: Deactivated successfully. Feb 13 15:37:39.652453 containerd[1992]: time="2025-02-13T15:37:39.652395610Z" level=error msg="Failed to destroy network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:39.655198 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800-shm.mount: Deactivated successfully. Feb 13 15:37:39.655979 containerd[1992]: time="2025-02-13T15:37:39.655338747Z" level=error msg="encountered an error cleaning up failed sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:39.655979 containerd[1992]: time="2025-02-13T15:37:39.655427606Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:39.656134 kubelet[2499]: E0213 15:37:39.655736 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:39.656134 kubelet[2499]: E0213 15:37:39.655796 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:39.656134 kubelet[2499]: E0213 15:37:39.655818 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:39.656293 kubelet[2499]: E0213 15:37:39.655886 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:40.309112 kubelet[2499]: E0213 15:37:40.308292 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:40.583013 kubelet[2499]: I0213 15:37:40.582772 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800" Feb 13 15:37:40.586480 containerd[1992]: time="2025-02-13T15:37:40.584949614Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:40.586480 containerd[1992]: time="2025-02-13T15:37:40.585364493Z" level=info msg="Ensure that sandbox 3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800 in task-service has been cleanup successfully" Feb 13 15:37:40.598302 containerd[1992]: time="2025-02-13T15:37:40.597629024Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:40.598302 containerd[1992]: time="2025-02-13T15:37:40.597677422Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:40.598865 systemd[1]: run-netns-cni\x2d32ac3bac\x2d5732\x2d7728\x2dde36\x2d2db7e44c7100.mount: Deactivated successfully. Feb 13 15:37:40.603169 containerd[1992]: time="2025-02-13T15:37:40.599720152Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:40.603169 containerd[1992]: time="2025-02-13T15:37:40.599894461Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:40.603169 containerd[1992]: time="2025-02-13T15:37:40.599912052Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:40.609280 containerd[1992]: time="2025-02-13T15:37:40.608510775Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:2,}" Feb 13 15:37:40.802224 containerd[1992]: time="2025-02-13T15:37:40.802175602Z" level=error msg="Failed to destroy network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:40.808446 containerd[1992]: time="2025-02-13T15:37:40.808396308Z" level=error msg="encountered an error cleaning up failed sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:40.808653 containerd[1992]: time="2025-02-13T15:37:40.808629943Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:40.809056 kubelet[2499]: E0213 15:37:40.809036 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:40.810219 kubelet[2499]: E0213 15:37:40.810198 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:40.810881 kubelet[2499]: E0213 15:37:40.810861 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:40.811808 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71-shm.mount: Deactivated successfully. Feb 13 15:37:40.814820 kubelet[2499]: E0213 15:37:40.814214 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:41.309517 kubelet[2499]: E0213 15:37:41.309482 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:41.434098 kubelet[2499]: I0213 15:37:41.432691 2499 topology_manager.go:215] "Topology Admit Handler" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" podNamespace="default" podName="nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:41.520823 kubelet[2499]: I0213 15:37:41.520784 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw9bl\" (UniqueName: \"kubernetes.io/projected/a90b5295-846c-4b90-98b2-8b1ab6c5419a-kube-api-access-gw9bl\") pod \"nginx-deployment-6d5f899847-hhn2p\" (UID: \"a90b5295-846c-4b90-98b2-8b1ab6c5419a\") " pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:41.587982 kubelet[2499]: I0213 15:37:41.587879 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71" Feb 13 15:37:41.590474 containerd[1992]: time="2025-02-13T15:37:41.590436790Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:41.592002 containerd[1992]: time="2025-02-13T15:37:41.591974896Z" level=info msg="Ensure that sandbox e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71 in task-service has been cleanup successfully" Feb 13 15:37:41.592285 containerd[1992]: time="2025-02-13T15:37:41.592266427Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:41.592400 containerd[1992]: time="2025-02-13T15:37:41.592384377Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:41.592896 containerd[1992]: time="2025-02-13T15:37:41.592853336Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:41.592970 containerd[1992]: time="2025-02-13T15:37:41.592954520Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:41.593024 containerd[1992]: time="2025-02-13T15:37:41.592972120Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:41.595603 containerd[1992]: time="2025-02-13T15:37:41.595501432Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:41.595603 containerd[1992]: time="2025-02-13T15:37:41.595599572Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:41.595723 containerd[1992]: time="2025-02-13T15:37:41.595613665Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:41.596550 systemd[1]: run-netns-cni\x2d23a47049\x2db90e\x2daf6c\x2d1bb9\x2d9663c9abd998.mount: Deactivated successfully. Feb 13 15:37:41.598737 containerd[1992]: time="2025-02-13T15:37:41.598701419Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:3,}" Feb 13 15:37:41.739614 containerd[1992]: time="2025-02-13T15:37:41.739569647Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:0,}" Feb 13 15:37:41.768544 containerd[1992]: time="2025-02-13T15:37:41.768363653Z" level=error msg="Failed to destroy network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.774785 containerd[1992]: time="2025-02-13T15:37:41.774736427Z" level=error msg="encountered an error cleaning up failed sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.774919 containerd[1992]: time="2025-02-13T15:37:41.774819177Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.775554 kubelet[2499]: E0213 15:37:41.775100 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.776111 kubelet[2499]: E0213 15:37:41.775260 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:41.776111 kubelet[2499]: E0213 15:37:41.775800 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:41.776111 kubelet[2499]: E0213 15:37:41.775886 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:41.780720 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe-shm.mount: Deactivated successfully. Feb 13 15:37:41.907902 containerd[1992]: time="2025-02-13T15:37:41.907746942Z" level=error msg="Failed to destroy network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.908636 containerd[1992]: time="2025-02-13T15:37:41.908499120Z" level=error msg="encountered an error cleaning up failed sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.909236 containerd[1992]: time="2025-02-13T15:37:41.908622756Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:0,} failed, error" error="failed to setup network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.910103 kubelet[2499]: E0213 15:37:41.909964 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:41.910103 kubelet[2499]: E0213 15:37:41.910047 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:41.910867 kubelet[2499]: E0213 15:37:41.910411 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:41.910983 kubelet[2499]: E0213 15:37:41.910965 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:42.150786 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Feb 13 15:37:42.310641 kubelet[2499]: E0213 15:37:42.310501 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:42.597866 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81-shm.mount: Deactivated successfully. Feb 13 15:37:42.605541 kubelet[2499]: I0213 15:37:42.604557 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe" Feb 13 15:37:42.605741 containerd[1992]: time="2025-02-13T15:37:42.605481019Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:42.606241 containerd[1992]: time="2025-02-13T15:37:42.605861415Z" level=info msg="Ensure that sandbox ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe in task-service has been cleanup successfully" Feb 13 15:37:42.608972 containerd[1992]: time="2025-02-13T15:37:42.608578244Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:42.608972 containerd[1992]: time="2025-02-13T15:37:42.608604469Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:42.610876 containerd[1992]: time="2025-02-13T15:37:42.610845408Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:42.611820 systemd[1]: run-netns-cni\x2d045116dc\x2d848b\x2df32b\x2df90d\x2dca889f6212a2.mount: Deactivated successfully. Feb 13 15:37:42.612256 containerd[1992]: time="2025-02-13T15:37:42.612040870Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:42.612256 containerd[1992]: time="2025-02-13T15:37:42.612219983Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613196333Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613540012Z" level=info msg="Ensure that sandbox 394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81 in task-service has been cleanup successfully" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613703332Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613781006Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613795145Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613876406Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:42.613979 containerd[1992]: time="2025-02-13T15:37:42.613889725Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:42.617761 kubelet[2499]: I0213 15:37:42.612588 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81" Feb 13 15:37:42.620278 systemd[1]: run-netns-cni\x2dca86a838\x2d8256\x2d2656\x2d5738\x2de07fe1a0d4f5.mount: Deactivated successfully. Feb 13 15:37:42.622096 containerd[1992]: time="2025-02-13T15:37:42.621961724Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:42.622286 containerd[1992]: time="2025-02-13T15:37:42.622265875Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:42.622533 containerd[1992]: time="2025-02-13T15:37:42.622375433Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:42.623002 containerd[1992]: time="2025-02-13T15:37:42.622326660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:1,}" Feb 13 15:37:42.624453 containerd[1992]: time="2025-02-13T15:37:42.624428592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:4,}" Feb 13 15:37:42.866118 containerd[1992]: time="2025-02-13T15:37:42.865449500Z" level=error msg="Failed to destroy network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.868788 containerd[1992]: time="2025-02-13T15:37:42.868737458Z" level=error msg="encountered an error cleaning up failed sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.868895 containerd[1992]: time="2025-02-13T15:37:42.868831485Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:1,} failed, error" error="failed to setup network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.870896 kubelet[2499]: E0213 15:37:42.869599 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.870896 kubelet[2499]: E0213 15:37:42.869697 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:42.870896 kubelet[2499]: E0213 15:37:42.869727 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:42.871139 kubelet[2499]: E0213 15:37:42.870752 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:42.873921 containerd[1992]: time="2025-02-13T15:37:42.873769673Z" level=error msg="Failed to destroy network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.875019 containerd[1992]: time="2025-02-13T15:37:42.874854874Z" level=error msg="encountered an error cleaning up failed sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.875179 containerd[1992]: time="2025-02-13T15:37:42.875035648Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.875316 kubelet[2499]: E0213 15:37:42.875289 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:42.875486 kubelet[2499]: E0213 15:37:42.875358 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:42.875486 kubelet[2499]: E0213 15:37:42.875387 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:42.875486 kubelet[2499]: E0213 15:37:42.875451 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:43.311661 kubelet[2499]: E0213 15:37:43.311620 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:43.597191 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d-shm.mount: Deactivated successfully. Feb 13 15:37:43.597386 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85-shm.mount: Deactivated successfully. Feb 13 15:37:43.620999 kubelet[2499]: I0213 15:37:43.620966 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d" Feb 13 15:37:43.621781 containerd[1992]: time="2025-02-13T15:37:43.621740128Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:43.624864 containerd[1992]: time="2025-02-13T15:37:43.622094855Z" level=info msg="Ensure that sandbox c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d in task-service has been cleanup successfully" Feb 13 15:37:43.624864 containerd[1992]: time="2025-02-13T15:37:43.622328774Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:43.624864 containerd[1992]: time="2025-02-13T15:37:43.622367962Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:43.624864 containerd[1992]: time="2025-02-13T15:37:43.624685461Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:43.625053 containerd[1992]: time="2025-02-13T15:37:43.624989763Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:43.625053 containerd[1992]: time="2025-02-13T15:37:43.625008209Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:43.626795 containerd[1992]: time="2025-02-13T15:37:43.625422769Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:43.626795 containerd[1992]: time="2025-02-13T15:37:43.625523439Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:43.626795 containerd[1992]: time="2025-02-13T15:37:43.625538394Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:43.626928 systemd[1]: run-netns-cni\x2d6b0261e2\x2dfd88\x2dc0d9\x2d2d9e\x2dad6fcac53817.mount: Deactivated successfully. Feb 13 15:37:43.630585 containerd[1992]: time="2025-02-13T15:37:43.630536513Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:43.630701 containerd[1992]: time="2025-02-13T15:37:43.630643212Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:43.630701 containerd[1992]: time="2025-02-13T15:37:43.630659027Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:43.631629 containerd[1992]: time="2025-02-13T15:37:43.631511517Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:43.631906 containerd[1992]: time="2025-02-13T15:37:43.631855616Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:43.631906 containerd[1992]: time="2025-02-13T15:37:43.631878316Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:43.633247 containerd[1992]: time="2025-02-13T15:37:43.632702623Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:5,}" Feb 13 15:37:43.633333 kubelet[2499]: I0213 15:37:43.632671 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85" Feb 13 15:37:43.634384 containerd[1992]: time="2025-02-13T15:37:43.634302444Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:43.634616 containerd[1992]: time="2025-02-13T15:37:43.634594955Z" level=info msg="Ensure that sandbox 2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85 in task-service has been cleanup successfully" Feb 13 15:37:43.634985 containerd[1992]: time="2025-02-13T15:37:43.634959768Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:43.635099 containerd[1992]: time="2025-02-13T15:37:43.634984931Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:43.638411 containerd[1992]: time="2025-02-13T15:37:43.638373251Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:43.638503 containerd[1992]: time="2025-02-13T15:37:43.638479314Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:43.638503 containerd[1992]: time="2025-02-13T15:37:43.638495768Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:43.639883 systemd[1]: run-netns-cni\x2d7432fe7c\x2d75f1\x2d1398\x2de41a\x2d88c93db0092d.mount: Deactivated successfully. Feb 13 15:37:43.641841 containerd[1992]: time="2025-02-13T15:37:43.641535280Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:2,}" Feb 13 15:37:43.830762 containerd[1992]: time="2025-02-13T15:37:43.830692821Z" level=error msg="Failed to destroy network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.833237 containerd[1992]: time="2025-02-13T15:37:43.833187132Z" level=error msg="encountered an error cleaning up failed sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.834043 containerd[1992]: time="2025-02-13T15:37:43.833430246Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:2,} failed, error" error="failed to setup network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.834315 kubelet[2499]: E0213 15:37:43.833675 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.834315 kubelet[2499]: E0213 15:37:43.833725 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:43.834315 kubelet[2499]: E0213 15:37:43.833756 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:43.834483 kubelet[2499]: E0213 15:37:43.833828 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:43.840605 containerd[1992]: time="2025-02-13T15:37:43.840195316Z" level=error msg="Failed to destroy network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.841034 containerd[1992]: time="2025-02-13T15:37:43.841000472Z" level=error msg="encountered an error cleaning up failed sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.841261 containerd[1992]: time="2025-02-13T15:37:43.841230686Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.842547 kubelet[2499]: E0213 15:37:43.842111 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:43.842547 kubelet[2499]: E0213 15:37:43.842182 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:43.842547 kubelet[2499]: E0213 15:37:43.842208 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:43.842683 kubelet[2499]: E0213 15:37:43.842260 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:44.283471 kubelet[2499]: E0213 15:37:44.283090 2499 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:44.313103 kubelet[2499]: E0213 15:37:44.313053 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:44.602005 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014-shm.mount: Deactivated successfully. Feb 13 15:37:44.641387 kubelet[2499]: I0213 15:37:44.640225 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014" Feb 13 15:37:44.641551 containerd[1992]: time="2025-02-13T15:37:44.641438066Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:37:44.644444 containerd[1992]: time="2025-02-13T15:37:44.641671042Z" level=info msg="Ensure that sandbox 4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014 in task-service has been cleanup successfully" Feb 13 15:37:44.644444 containerd[1992]: time="2025-02-13T15:37:44.642116031Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:37:44.644444 containerd[1992]: time="2025-02-13T15:37:44.642136001Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:37:44.645363 containerd[1992]: time="2025-02-13T15:37:44.644986517Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:44.645475 systemd[1]: run-netns-cni\x2dbe5488b0\x2d726a\x2d2bde\x2d4c60\x2d51e35f4f8daa.mount: Deactivated successfully. Feb 13 15:37:44.647704 containerd[1992]: time="2025-02-13T15:37:44.647672346Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:44.647704 containerd[1992]: time="2025-02-13T15:37:44.647703151Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:44.649312 containerd[1992]: time="2025-02-13T15:37:44.648960687Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:44.650292 containerd[1992]: time="2025-02-13T15:37:44.650267072Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:44.650292 containerd[1992]: time="2025-02-13T15:37:44.650287873Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:44.651055 containerd[1992]: time="2025-02-13T15:37:44.650844639Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:44.651055 containerd[1992]: time="2025-02-13T15:37:44.650962192Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:44.651055 containerd[1992]: time="2025-02-13T15:37:44.650979010Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:44.652052 kubelet[2499]: I0213 15:37:44.651586 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab" Feb 13 15:37:44.652158 containerd[1992]: time="2025-02-13T15:37:44.651661382Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:44.652158 containerd[1992]: time="2025-02-13T15:37:44.651766683Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:44.652158 containerd[1992]: time="2025-02-13T15:37:44.651782540Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:44.652158 containerd[1992]: time="2025-02-13T15:37:44.652083239Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:37:44.652376 containerd[1992]: time="2025-02-13T15:37:44.652330614Z" level=info msg="Ensure that sandbox dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab in task-service has been cleanup successfully" Feb 13 15:37:44.655088 containerd[1992]: time="2025-02-13T15:37:44.653120734Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:37:44.655088 containerd[1992]: time="2025-02-13T15:37:44.653143284Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:37:44.655088 containerd[1992]: time="2025-02-13T15:37:44.653199709Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:44.655267 containerd[1992]: time="2025-02-13T15:37:44.653456469Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:44.655267 containerd[1992]: time="2025-02-13T15:37:44.655129179Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:44.657984 systemd[1]: run-netns-cni\x2d2cfb7f5d\x2da021\x2dad2b\x2ddf5e\x2da70036dd39b4.mount: Deactivated successfully. Feb 13 15:37:44.660234 containerd[1992]: time="2025-02-13T15:37:44.660154102Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:44.660367 containerd[1992]: time="2025-02-13T15:37:44.660281326Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:44.660367 containerd[1992]: time="2025-02-13T15:37:44.660296321Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:44.660452 containerd[1992]: time="2025-02-13T15:37:44.660435837Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:6,}" Feb 13 15:37:44.662122 containerd[1992]: time="2025-02-13T15:37:44.661389433Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:44.662122 containerd[1992]: time="2025-02-13T15:37:44.661723810Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:44.662122 containerd[1992]: time="2025-02-13T15:37:44.661746507Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:44.663009 containerd[1992]: time="2025-02-13T15:37:44.662983762Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:3,}" Feb 13 15:37:44.898181 containerd[1992]: time="2025-02-13T15:37:44.898116254Z" level=error msg="Failed to destroy network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.898834 containerd[1992]: time="2025-02-13T15:37:44.898771858Z" level=error msg="encountered an error cleaning up failed sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.899119 containerd[1992]: time="2025-02-13T15:37:44.898973291Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:3,} failed, error" error="failed to setup network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.900570 kubelet[2499]: E0213 15:37:44.900548 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.900818 kubelet[2499]: E0213 15:37:44.900804 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:44.900922 kubelet[2499]: E0213 15:37:44.900913 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:44.901260 kubelet[2499]: E0213 15:37:44.901244 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:44.909707 containerd[1992]: time="2025-02-13T15:37:44.908992072Z" level=error msg="Failed to destroy network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.910503 containerd[1992]: time="2025-02-13T15:37:44.910464962Z" level=error msg="encountered an error cleaning up failed sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.910621 containerd[1992]: time="2025-02-13T15:37:44.910538979Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.911320 kubelet[2499]: E0213 15:37:44.911123 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:44.911320 kubelet[2499]: E0213 15:37:44.911200 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:44.911320 kubelet[2499]: E0213 15:37:44.911290 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:44.913634 kubelet[2499]: E0213 15:37:44.913429 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:45.315725 kubelet[2499]: E0213 15:37:45.313266 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:45.611891 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93-shm.mount: Deactivated successfully. Feb 13 15:37:45.666666 kubelet[2499]: I0213 15:37:45.666559 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93" Feb 13 15:37:45.669338 containerd[1992]: time="2025-02-13T15:37:45.668705546Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:37:45.669338 containerd[1992]: time="2025-02-13T15:37:45.669126305Z" level=info msg="Ensure that sandbox 3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93 in task-service has been cleanup successfully" Feb 13 15:37:45.675154 containerd[1992]: time="2025-02-13T15:37:45.674744818Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:37:45.675154 containerd[1992]: time="2025-02-13T15:37:45.674872850Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:37:45.676582 systemd[1]: run-netns-cni\x2de159658e\x2d3ce2\x2d29b8\x2d249a\x2d9bb0d0d58b96.mount: Deactivated successfully. Feb 13 15:37:45.680215 containerd[1992]: time="2025-02-13T15:37:45.680180545Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:37:45.680677 containerd[1992]: time="2025-02-13T15:37:45.680656582Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:37:45.680783 containerd[1992]: time="2025-02-13T15:37:45.680759478Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:37:45.681563 containerd[1992]: time="2025-02-13T15:37:45.681535896Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:45.681877 containerd[1992]: time="2025-02-13T15:37:45.681633240Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:45.681877 containerd[1992]: time="2025-02-13T15:37:45.681650877Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:45.682249 kubelet[2499]: I0213 15:37:45.682171 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689" Feb 13 15:37:45.683688 containerd[1992]: time="2025-02-13T15:37:45.683045548Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:45.683688 containerd[1992]: time="2025-02-13T15:37:45.683620052Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:45.683890 containerd[1992]: time="2025-02-13T15:37:45.683685197Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:45.685629 containerd[1992]: time="2025-02-13T15:37:45.684901905Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:45.685629 containerd[1992]: time="2025-02-13T15:37:45.685331918Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:45.685629 containerd[1992]: time="2025-02-13T15:37:45.685349335Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:45.686092 containerd[1992]: time="2025-02-13T15:37:45.685973226Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:37:45.686672 containerd[1992]: time="2025-02-13T15:37:45.686528060Z" level=info msg="Ensure that sandbox 6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689 in task-service has been cleanup successfully" Feb 13 15:37:45.686975 containerd[1992]: time="2025-02-13T15:37:45.686934885Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:45.687051 containerd[1992]: time="2025-02-13T15:37:45.687025568Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:45.687051 containerd[1992]: time="2025-02-13T15:37:45.687041850Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:45.694906 containerd[1992]: time="2025-02-13T15:37:45.691555141Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:45.694906 containerd[1992]: time="2025-02-13T15:37:45.691675250Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:45.694906 containerd[1992]: time="2025-02-13T15:37:45.691689996Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:45.694906 containerd[1992]: time="2025-02-13T15:37:45.691839769Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:37:45.694906 containerd[1992]: time="2025-02-13T15:37:45.691858078Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:37:45.694206 systemd[1]: run-netns-cni\x2de9314e47\x2d3da4\x2d7190\x2d1b00\x2dafaf4f79d20c.mount: Deactivated successfully. Feb 13 15:37:45.696349 containerd[1992]: time="2025-02-13T15:37:45.696250795Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:7,}" Feb 13 15:37:45.697787 containerd[1992]: time="2025-02-13T15:37:45.696257789Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:37:45.697939 containerd[1992]: time="2025-02-13T15:37:45.697882460Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:37:45.697939 containerd[1992]: time="2025-02-13T15:37:45.697901573Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:37:45.698929 containerd[1992]: time="2025-02-13T15:37:45.698898534Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:45.699017 containerd[1992]: time="2025-02-13T15:37:45.698996404Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:45.699017 containerd[1992]: time="2025-02-13T15:37:45.699012934Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:45.704830 containerd[1992]: time="2025-02-13T15:37:45.704547188Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:45.704830 containerd[1992]: time="2025-02-13T15:37:45.704764148Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:45.704830 containerd[1992]: time="2025-02-13T15:37:45.704782432Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:45.709805 containerd[1992]: time="2025-02-13T15:37:45.709749817Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:4,}" Feb 13 15:37:46.063395 containerd[1992]: time="2025-02-13T15:37:46.063342795Z" level=error msg="Failed to destroy network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.065248 containerd[1992]: time="2025-02-13T15:37:46.065198536Z" level=error msg="Failed to destroy network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.065603 containerd[1992]: time="2025-02-13T15:37:46.065572034Z" level=error msg="encountered an error cleaning up failed sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.065738 containerd[1992]: time="2025-02-13T15:37:46.065708112Z" level=error msg="encountered an error cleaning up failed sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.065795 containerd[1992]: time="2025-02-13T15:37:46.065765579Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:7,} failed, error" error="failed to setup network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.066124 containerd[1992]: time="2025-02-13T15:37:46.065858599Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:4,} failed, error" error="failed to setup network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.066257 kubelet[2499]: E0213 15:37:46.066021 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.066257 kubelet[2499]: E0213 15:37:46.066162 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:46.066257 kubelet[2499]: E0213 15:37:46.066194 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:46.066894 kubelet[2499]: E0213 15:37:46.066873 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:46.067040 kubelet[2499]: E0213 15:37:46.067024 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.067118 kubelet[2499]: E0213 15:37:46.067093 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:46.067177 kubelet[2499]: E0213 15:37:46.067123 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:46.067224 kubelet[2499]: E0213 15:37:46.067180 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:46.315628 kubelet[2499]: E0213 15:37:46.314166 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:46.602177 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6-shm.mount: Deactivated successfully. Feb 13 15:37:46.603126 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf-shm.mount: Deactivated successfully. Feb 13 15:37:46.696816 kubelet[2499]: I0213 15:37:46.696782 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf" Feb 13 15:37:46.701157 containerd[1992]: time="2025-02-13T15:37:46.698692185Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:37:46.701157 containerd[1992]: time="2025-02-13T15:37:46.699032197Z" level=info msg="Ensure that sandbox 88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf in task-service has been cleanup successfully" Feb 13 15:37:46.702499 containerd[1992]: time="2025-02-13T15:37:46.702397551Z" level=info msg="TearDown network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" successfully" Feb 13 15:37:46.702499 containerd[1992]: time="2025-02-13T15:37:46.702424901Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" returns successfully" Feb 13 15:37:46.705318 containerd[1992]: time="2025-02-13T15:37:46.702894141Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:37:46.705566 containerd[1992]: time="2025-02-13T15:37:46.705490828Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:37:46.705772 containerd[1992]: time="2025-02-13T15:37:46.705706464Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:37:46.706009 systemd[1]: run-netns-cni\x2d0e487226\x2db884\x2dce61\x2da454\x2d89efbcb40834.mount: Deactivated successfully. Feb 13 15:37:46.708677 containerd[1992]: time="2025-02-13T15:37:46.707672859Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:37:46.708677 containerd[1992]: time="2025-02-13T15:37:46.707856834Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:37:46.708677 containerd[1992]: time="2025-02-13T15:37:46.707876030Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:37:46.708987 containerd[1992]: time="2025-02-13T15:37:46.708693093Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:46.708987 containerd[1992]: time="2025-02-13T15:37:46.708919873Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:46.708987 containerd[1992]: time="2025-02-13T15:37:46.708938459Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:46.713925 containerd[1992]: time="2025-02-13T15:37:46.713630175Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:46.713925 containerd[1992]: time="2025-02-13T15:37:46.713729810Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:46.713925 containerd[1992]: time="2025-02-13T15:37:46.713746699Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:46.715869 containerd[1992]: time="2025-02-13T15:37:46.714369348Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:46.716171 containerd[1992]: time="2025-02-13T15:37:46.716133329Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:46.716390 containerd[1992]: time="2025-02-13T15:37:46.716301103Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:46.716762 kubelet[2499]: I0213 15:37:46.716737 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.717693027Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.717803633Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.717834283Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.718421524Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.718514117Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:46.718735 containerd[1992]: time="2025-02-13T15:37:46.718528878Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:46.719176 containerd[1992]: time="2025-02-13T15:37:46.719125205Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:37:46.719540 containerd[1992]: time="2025-02-13T15:37:46.719513861Z" level=info msg="Ensure that sandbox 3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6 in task-service has been cleanup successfully" Feb 13 15:37:46.719835 containerd[1992]: time="2025-02-13T15:37:46.719812783Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:8,}" Feb 13 15:37:46.723266 containerd[1992]: time="2025-02-13T15:37:46.723230699Z" level=info msg="TearDown network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" successfully" Feb 13 15:37:46.723266 containerd[1992]: time="2025-02-13T15:37:46.723265629Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" returns successfully" Feb 13 15:37:46.724996 containerd[1992]: time="2025-02-13T15:37:46.724177397Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:37:46.724996 containerd[1992]: time="2025-02-13T15:37:46.724291493Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:37:46.724996 containerd[1992]: time="2025-02-13T15:37:46.724307594Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:37:46.726028 systemd[1]: run-netns-cni\x2d1602d8a2\x2d9a84\x2d208c\x2dafe6\x2d1256cf5af339.mount: Deactivated successfully. Feb 13 15:37:46.731776 containerd[1992]: time="2025-02-13T15:37:46.730486936Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:37:46.731776 containerd[1992]: time="2025-02-13T15:37:46.731030025Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:37:46.731776 containerd[1992]: time="2025-02-13T15:37:46.731051038Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:37:46.734172 containerd[1992]: time="2025-02-13T15:37:46.734087744Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:46.734690 containerd[1992]: time="2025-02-13T15:37:46.734615305Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:46.734690 containerd[1992]: time="2025-02-13T15:37:46.734636158Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:46.735471 containerd[1992]: time="2025-02-13T15:37:46.735442229Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:46.735682 containerd[1992]: time="2025-02-13T15:37:46.735647255Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:46.735682 containerd[1992]: time="2025-02-13T15:37:46.735666903Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:46.737712 containerd[1992]: time="2025-02-13T15:37:46.737684715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:5,}" Feb 13 15:37:46.938513 containerd[1992]: time="2025-02-13T15:37:46.938460830Z" level=error msg="Failed to destroy network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.938841 containerd[1992]: time="2025-02-13T15:37:46.938808031Z" level=error msg="encountered an error cleaning up failed sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.938921 containerd[1992]: time="2025-02-13T15:37:46.938892598Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:5,} failed, error" error="failed to setup network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.939212 kubelet[2499]: E0213 15:37:46.939188 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.939305 kubelet[2499]: E0213 15:37:46.939254 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:46.939305 kubelet[2499]: E0213 15:37:46.939286 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:46.939400 kubelet[2499]: E0213 15:37:46.939357 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:46.966634 containerd[1992]: time="2025-02-13T15:37:46.966426207Z" level=error msg="Failed to destroy network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.966941 containerd[1992]: time="2025-02-13T15:37:46.966906511Z" level=error msg="encountered an error cleaning up failed sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.967222 containerd[1992]: time="2025-02-13T15:37:46.966975588Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:8,} failed, error" error="failed to setup network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.967913 kubelet[2499]: E0213 15:37:46.967497 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:46.967913 kubelet[2499]: E0213 15:37:46.967562 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:46.967913 kubelet[2499]: E0213 15:37:46.967592 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:46.968054 kubelet[2499]: E0213 15:37:46.967663 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:47.315132 kubelet[2499]: E0213 15:37:47.314990 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:47.542693 containerd[1992]: time="2025-02-13T15:37:47.542631805Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:47.548600 containerd[1992]: time="2025-02-13T15:37:47.548371512Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Feb 13 15:37:47.553101 containerd[1992]: time="2025-02-13T15:37:47.553016544Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:47.557985 containerd[1992]: time="2025-02-13T15:37:47.557911333Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:47.558999 containerd[1992]: time="2025-02-13T15:37:47.558892817Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 8.980192017s" Feb 13 15:37:47.558999 containerd[1992]: time="2025-02-13T15:37:47.558994854Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Feb 13 15:37:47.589713 containerd[1992]: time="2025-02-13T15:37:47.589591893Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Feb 13 15:37:47.610848 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7-shm.mount: Deactivated successfully. Feb 13 15:37:47.611856 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47-shm.mount: Deactivated successfully. Feb 13 15:37:47.613977 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1233480605.mount: Deactivated successfully. Feb 13 15:37:47.665053 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3212982959.mount: Deactivated successfully. Feb 13 15:37:47.675944 containerd[1992]: time="2025-02-13T15:37:47.675372107Z" level=info msg="CreateContainer within sandbox \"5911a437b90c9679f7bbd888cd2f444c0f5ce01d36f6f4d3f31669bc53e02ca0\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"269c2a58ea1db8b5af50faf96731d848b3972b85407c4a8066160f390784307e\"" Feb 13 15:37:47.678198 containerd[1992]: time="2025-02-13T15:37:47.676835477Z" level=info msg="StartContainer for \"269c2a58ea1db8b5af50faf96731d848b3972b85407c4a8066160f390784307e\"" Feb 13 15:37:47.745261 kubelet[2499]: I0213 15:37:47.744549 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47" Feb 13 15:37:47.746742 containerd[1992]: time="2025-02-13T15:37:47.746019491Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" Feb 13 15:37:47.746742 containerd[1992]: time="2025-02-13T15:37:47.746521791Z" level=info msg="Ensure that sandbox c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47 in task-service has been cleanup successfully" Feb 13 15:37:47.748022 containerd[1992]: time="2025-02-13T15:37:47.747617658Z" level=info msg="TearDown network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" successfully" Feb 13 15:37:47.748022 containerd[1992]: time="2025-02-13T15:37:47.747654424Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" returns successfully" Feb 13 15:37:47.749031 containerd[1992]: time="2025-02-13T15:37:47.748733207Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:37:47.749031 containerd[1992]: time="2025-02-13T15:37:47.748855057Z" level=info msg="TearDown network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" successfully" Feb 13 15:37:47.749031 containerd[1992]: time="2025-02-13T15:37:47.748931060Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" returns successfully" Feb 13 15:37:47.750143 containerd[1992]: time="2025-02-13T15:37:47.749916271Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:37:47.750466 containerd[1992]: time="2025-02-13T15:37:47.750336171Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:37:47.750466 containerd[1992]: time="2025-02-13T15:37:47.750356154Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:37:47.755321 containerd[1992]: time="2025-02-13T15:37:47.754905028Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:37:47.755321 containerd[1992]: time="2025-02-13T15:37:47.755020419Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:37:47.755321 containerd[1992]: time="2025-02-13T15:37:47.755034339Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:37:47.755696 containerd[1992]: time="2025-02-13T15:37:47.755666077Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:47.756065 containerd[1992]: time="2025-02-13T15:37:47.755777200Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:47.756065 containerd[1992]: time="2025-02-13T15:37:47.755869720Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:47.763653 systemd[1]: run-netns-cni\x2d9df706db\x2d8df1\x2d72cb\x2d1318\x2dc80a88e6266c.mount: Deactivated successfully. Feb 13 15:37:47.766096 containerd[1992]: time="2025-02-13T15:37:47.764768255Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:47.766096 containerd[1992]: time="2025-02-13T15:37:47.764891012Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:47.766096 containerd[1992]: time="2025-02-13T15:37:47.764906011Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:47.769814 containerd[1992]: time="2025-02-13T15:37:47.769093433Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:47.769814 containerd[1992]: time="2025-02-13T15:37:47.769212752Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:47.769814 containerd[1992]: time="2025-02-13T15:37:47.769227884Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:47.770523 containerd[1992]: time="2025-02-13T15:37:47.770226235Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:47.770523 containerd[1992]: time="2025-02-13T15:37:47.770321675Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:47.770523 containerd[1992]: time="2025-02-13T15:37:47.770337205Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:47.770928 kubelet[2499]: I0213 15:37:47.770907 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7" Feb 13 15:37:47.771064 containerd[1992]: time="2025-02-13T15:37:47.771044608Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:47.771239 containerd[1992]: time="2025-02-13T15:37:47.771222314Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:47.771325 containerd[1992]: time="2025-02-13T15:37:47.771310427Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:47.772331 containerd[1992]: time="2025-02-13T15:37:47.772309665Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" Feb 13 15:37:47.773280 containerd[1992]: time="2025-02-13T15:37:47.773129326Z" level=info msg="Ensure that sandbox af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7 in task-service has been cleanup successfully" Feb 13 15:37:47.773401 containerd[1992]: time="2025-02-13T15:37:47.772643360Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:9,}" Feb 13 15:37:47.774025 containerd[1992]: time="2025-02-13T15:37:47.773987217Z" level=info msg="TearDown network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" successfully" Feb 13 15:37:47.774287 containerd[1992]: time="2025-02-13T15:37:47.774214712Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" returns successfully" Feb 13 15:37:47.774681 containerd[1992]: time="2025-02-13T15:37:47.774654035Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:37:47.774779 containerd[1992]: time="2025-02-13T15:37:47.774745717Z" level=info msg="TearDown network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" successfully" Feb 13 15:37:47.774779 containerd[1992]: time="2025-02-13T15:37:47.774760836Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" returns successfully" Feb 13 15:37:47.792777 containerd[1992]: time="2025-02-13T15:37:47.791355027Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:37:47.792777 containerd[1992]: time="2025-02-13T15:37:47.791536549Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:37:47.792777 containerd[1992]: time="2025-02-13T15:37:47.791571767Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:37:47.792997 containerd[1992]: time="2025-02-13T15:37:47.792851232Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:37:47.792997 containerd[1992]: time="2025-02-13T15:37:47.792973500Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:37:47.792997 containerd[1992]: time="2025-02-13T15:37:47.792989366Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:37:47.793432 containerd[1992]: time="2025-02-13T15:37:47.793407103Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:47.793912 containerd[1992]: time="2025-02-13T15:37:47.793884985Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:47.793912 containerd[1992]: time="2025-02-13T15:37:47.793908432Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:47.794586 containerd[1992]: time="2025-02-13T15:37:47.794320085Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:47.795351 containerd[1992]: time="2025-02-13T15:37:47.795085570Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:47.795351 containerd[1992]: time="2025-02-13T15:37:47.795303981Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:47.800013 containerd[1992]: time="2025-02-13T15:37:47.799888330Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:6,}" Feb 13 15:37:47.944767 containerd[1992]: time="2025-02-13T15:37:47.944672121Z" level=info msg="StartContainer for \"269c2a58ea1db8b5af50faf96731d848b3972b85407c4a8066160f390784307e\" returns successfully" Feb 13 15:37:48.119616 containerd[1992]: time="2025-02-13T15:37:48.119565032Z" level=error msg="Failed to destroy network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.122960 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Feb 13 15:37:48.123252 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Feb 13 15:37:48.123503 containerd[1992]: time="2025-02-13T15:37:48.120503542Z" level=error msg="encountered an error cleaning up failed sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.124093 containerd[1992]: time="2025-02-13T15:37:48.124031126Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:9,} failed, error" error="failed to setup network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.126227 kubelet[2499]: E0213 15:37:48.126202 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.126499 kubelet[2499]: E0213 15:37:48.126484 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:48.126663 kubelet[2499]: E0213 15:37:48.126652 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-kxrfd" Feb 13 15:37:48.126884 kubelet[2499]: E0213 15:37:48.126852 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-kxrfd_calico-system(9afcbeb8-f8e4-4f2f-9b35-6cef169285f0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-kxrfd" podUID="9afcbeb8-f8e4-4f2f-9b35-6cef169285f0" Feb 13 15:37:48.134865 containerd[1992]: time="2025-02-13T15:37:48.134808391Z" level=error msg="Failed to destroy network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.135228 containerd[1992]: time="2025-02-13T15:37:48.135193800Z" level=error msg="encountered an error cleaning up failed sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.135347 containerd[1992]: time="2025-02-13T15:37:48.135264690Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:6,} failed, error" error="failed to setup network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.135563 kubelet[2499]: E0213 15:37:48.135515 2499 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 15:37:48.135659 kubelet[2499]: E0213 15:37:48.135574 2499 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:48.135659 kubelet[2499]: E0213 15:37:48.135604 2499 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-6d5f899847-hhn2p" Feb 13 15:37:48.135971 kubelet[2499]: E0213 15:37:48.135678 2499 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-6d5f899847-hhn2p_default(a90b5295-846c-4b90-98b2-8b1ab6c5419a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-6d5f899847-hhn2p" podUID="a90b5295-846c-4b90-98b2-8b1ab6c5419a" Feb 13 15:37:48.317062 kubelet[2499]: E0213 15:37:48.315809 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:48.603533 systemd[1]: run-netns-cni\x2d98463cfe\x2d13cc\x2d4bb2\x2d75ec\x2d776ce4f41f2d.mount: Deactivated successfully. Feb 13 15:37:48.786742 kubelet[2499]: I0213 15:37:48.786713 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8" Feb 13 15:37:48.801349 containerd[1992]: time="2025-02-13T15:37:48.796588988Z" level=info msg="StopPodSandbox for \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\"" Feb 13 15:37:48.801349 containerd[1992]: time="2025-02-13T15:37:48.796893442Z" level=info msg="Ensure that sandbox 9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8 in task-service has been cleanup successfully" Feb 13 15:37:48.811495 containerd[1992]: time="2025-02-13T15:37:48.809874806Z" level=info msg="TearDown network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" successfully" Feb 13 15:37:48.811495 containerd[1992]: time="2025-02-13T15:37:48.809924404Z" level=info msg="StopPodSandbox for \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" returns successfully" Feb 13 15:37:48.815614 systemd[1]: run-netns-cni\x2de76e3b46\x2d45aa\x2d98c6\x2ddf85\x2d7e1b32e0b85e.mount: Deactivated successfully. Feb 13 15:37:48.823597 containerd[1992]: time="2025-02-13T15:37:48.821336309Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" Feb 13 15:37:48.824420 kubelet[2499]: I0213 15:37:48.821636 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-node-4hcb2" podStartSLOduration=4.987588375 podStartE2EDuration="24.821569015s" podCreationTimestamp="2025-02-13 15:37:24 +0000 UTC" firstStartedPulling="2025-02-13 15:37:27.725695466 +0000 UTC m=+3.769340964" lastFinishedPulling="2025-02-13 15:37:47.559676108 +0000 UTC m=+23.603321604" observedRunningTime="2025-02-13 15:37:48.821037802 +0000 UTC m=+24.864683301" watchObservedRunningTime="2025-02-13 15:37:48.821569015 +0000 UTC m=+24.865214540" Feb 13 15:37:48.827118 containerd[1992]: time="2025-02-13T15:37:48.826199366Z" level=info msg="TearDown network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" successfully" Feb 13 15:37:48.827118 containerd[1992]: time="2025-02-13T15:37:48.826229588Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" returns successfully" Feb 13 15:37:48.827118 containerd[1992]: time="2025-02-13T15:37:48.827005947Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:37:48.827766 containerd[1992]: time="2025-02-13T15:37:48.827125496Z" level=info msg="TearDown network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" successfully" Feb 13 15:37:48.827766 containerd[1992]: time="2025-02-13T15:37:48.827141434Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" returns successfully" Feb 13 15:37:48.830940 containerd[1992]: time="2025-02-13T15:37:48.830895124Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:37:48.832372 containerd[1992]: time="2025-02-13T15:37:48.831001204Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:37:48.832372 containerd[1992]: time="2025-02-13T15:37:48.831017616Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:37:48.840554 containerd[1992]: time="2025-02-13T15:37:48.840329670Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:37:48.840706 containerd[1992]: time="2025-02-13T15:37:48.840637428Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:37:48.840706 containerd[1992]: time="2025-02-13T15:37:48.840654732Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:37:48.841385 containerd[1992]: time="2025-02-13T15:37:48.841353734Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:37:48.841788 containerd[1992]: time="2025-02-13T15:37:48.841560384Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:37:48.841788 containerd[1992]: time="2025-02-13T15:37:48.841580814Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:37:48.842025 kubelet[2499]: I0213 15:37:48.841902 2499 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476" Feb 13 15:37:48.843254 containerd[1992]: time="2025-02-13T15:37:48.843048483Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:37:48.843254 containerd[1992]: time="2025-02-13T15:37:48.843182397Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:37:48.843254 containerd[1992]: time="2025-02-13T15:37:48.843198474Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:37:48.845925 containerd[1992]: time="2025-02-13T15:37:48.844900492Z" level=info msg="StopPodSandbox for \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\"" Feb 13 15:37:48.845925 containerd[1992]: time="2025-02-13T15:37:48.844966273Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:37:48.845925 containerd[1992]: time="2025-02-13T15:37:48.845719378Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:37:48.845925 containerd[1992]: time="2025-02-13T15:37:48.845918629Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:37:48.851715 containerd[1992]: time="2025-02-13T15:37:48.846812981Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:37:48.851715 containerd[1992]: time="2025-02-13T15:37:48.851595120Z" level=info msg="Ensure that sandbox b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476 in task-service has been cleanup successfully" Feb 13 15:37:48.852818 containerd[1992]: time="2025-02-13T15:37:48.852430139Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:37:48.852818 containerd[1992]: time="2025-02-13T15:37:48.852457898Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:37:48.864740 containerd[1992]: time="2025-02-13T15:37:48.852747030Z" level=info msg="TearDown network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" successfully" Feb 13 15:37:48.864740 containerd[1992]: time="2025-02-13T15:37:48.852884406Z" level=info msg="StopPodSandbox for \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" returns successfully" Feb 13 15:37:48.864740 containerd[1992]: time="2025-02-13T15:37:48.853015872Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:37:48.864740 containerd[1992]: time="2025-02-13T15:37:48.853123011Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:37:48.864740 containerd[1992]: time="2025-02-13T15:37:48.853136966Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:37:48.866623 systemd[1]: run-netns-cni\x2d4bafda53\x2dcd85\x2de3a9\x2dd079\x2d8a3d9bbfb3ec.mount: Deactivated successfully. Feb 13 15:37:48.869706 containerd[1992]: time="2025-02-13T15:37:48.867097663Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" Feb 13 15:37:48.869706 containerd[1992]: time="2025-02-13T15:37:48.867257627Z" level=info msg="TearDown network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" successfully" Feb 13 15:37:48.869706 containerd[1992]: time="2025-02-13T15:37:48.867271591Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" returns successfully" Feb 13 15:37:48.869706 containerd[1992]: time="2025-02-13T15:37:48.868106116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:10,}" Feb 13 15:37:48.871328 containerd[1992]: time="2025-02-13T15:37:48.871119935Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:37:48.871328 containerd[1992]: time="2025-02-13T15:37:48.871234420Z" level=info msg="TearDown network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" successfully" Feb 13 15:37:48.871328 containerd[1992]: time="2025-02-13T15:37:48.871246055Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" returns successfully" Feb 13 15:37:48.872572 containerd[1992]: time="2025-02-13T15:37:48.872542445Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:37:48.873007 containerd[1992]: time="2025-02-13T15:37:48.872977535Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:37:48.873181 containerd[1992]: time="2025-02-13T15:37:48.873004245Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:37:48.880115 containerd[1992]: time="2025-02-13T15:37:48.877830866Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:37:48.880115 containerd[1992]: time="2025-02-13T15:37:48.877961812Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:37:48.880115 containerd[1992]: time="2025-02-13T15:37:48.878030973Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:37:48.881831 containerd[1992]: time="2025-02-13T15:37:48.881790684Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:37:48.882215 containerd[1992]: time="2025-02-13T15:37:48.882193926Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:37:48.882291 containerd[1992]: time="2025-02-13T15:37:48.882219092Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:37:48.885096 containerd[1992]: time="2025-02-13T15:37:48.885045939Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:37:48.885194 containerd[1992]: time="2025-02-13T15:37:48.885180542Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:37:48.885250 containerd[1992]: time="2025-02-13T15:37:48.885197813Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:37:48.886746 containerd[1992]: time="2025-02-13T15:37:48.886388905Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:7,}" Feb 13 15:37:49.274646 (udev-worker)[3631]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:37:49.276505 systemd-networkd[1568]: cali53e2623664b: Link UP Feb 13 15:37:49.277485 systemd-networkd[1568]: cali53e2623664b: Gained carrier Feb 13 15:37:49.316583 kubelet[2499]: E0213 15:37:49.316536 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:48.988 [INFO][3594] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.055 [INFO][3594] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.20.193-k8s-csi--node--driver--kxrfd-eth0 csi-node-driver- calico-system 9afcbeb8-f8e4-4f2f-9b35-6cef169285f0 910 0 2025-02-13 15:37:24 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b695c467 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172.31.20.193 csi-node-driver-kxrfd eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali53e2623664b [] []}} ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.055 [INFO][3594] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.156 [INFO][3617] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" HandleID="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Workload="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.186 [INFO][3617] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" HandleID="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Workload="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003e2b90), Attrs:map[string]string{"namespace":"calico-system", "node":"172.31.20.193", "pod":"csi-node-driver-kxrfd", "timestamp":"2025-02-13 15:37:49.156039035 +0000 UTC"}, Hostname:"172.31.20.193", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.186 [INFO][3617] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.186 [INFO][3617] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.186 [INFO][3617] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.20.193' Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.189 [INFO][3617] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.197 [INFO][3617] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.211 [INFO][3617] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.216 [INFO][3617] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.222 [INFO][3617] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.222 [INFO][3617] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.229 [INFO][3617] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5 Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.243 [INFO][3617] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3617] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.65/26] block=192.168.106.64/26 handle="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3617] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.65/26] handle="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" host="172.31.20.193" Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3617] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 15:37:49.318060 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3617] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.65/26] IPv6=[] ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" HandleID="k8s-pod-network.394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Workload="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.261 [INFO][3594] cni-plugin/k8s.go 386: Populated endpoint ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-csi--node--driver--kxrfd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0", ResourceVersion:"910", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 37, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"", Pod:"csi-node-driver-kxrfd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.106.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali53e2623664b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.261 [INFO][3594] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.65/32] ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.261 [INFO][3594] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali53e2623664b ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.278 [INFO][3594] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.278 [INFO][3594] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-csi--node--driver--kxrfd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"9afcbeb8-f8e4-4f2f-9b35-6cef169285f0", ResourceVersion:"910", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 37, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b695c467", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5", Pod:"csi-node-driver-kxrfd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.106.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali53e2623664b", MAC:"0e:1f:56:64:d5:23", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:37:49.319829 containerd[1992]: 2025-02-13 15:37:49.316 [INFO][3594] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5" Namespace="calico-system" Pod="csi-node-driver-kxrfd" WorkloadEndpoint="172.31.20.193-k8s-csi--node--driver--kxrfd-eth0" Feb 13 15:37:49.355961 containerd[1992]: time="2025-02-13T15:37:49.355350132Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:37:49.355961 containerd[1992]: time="2025-02-13T15:37:49.355421463Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:37:49.355961 containerd[1992]: time="2025-02-13T15:37:49.355443100Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:49.355961 containerd[1992]: time="2025-02-13T15:37:49.355550472Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:49.356893 (udev-worker)[3568]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:37:49.361818 systemd-networkd[1568]: cali366bf208cdf: Link UP Feb 13 15:37:49.362921 systemd-networkd[1568]: cali366bf208cdf: Gained carrier Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:48.990 [INFO][3599] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.055 [INFO][3599] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0 nginx-deployment-6d5f899847- default a90b5295-846c-4b90-98b2-8b1ab6c5419a 1049 0 2025-02-13 15:37:41 +0000 UTC map[app:nginx pod-template-hash:6d5f899847 projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.31.20.193 nginx-deployment-6d5f899847-hhn2p eth0 default [] [] [kns.default ksa.default.default] cali366bf208cdf [] []}} ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.055 [INFO][3599] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.162 [INFO][3616] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" HandleID="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Workload="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.192 [INFO][3616] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" HandleID="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Workload="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000050ad0), Attrs:map[string]string{"namespace":"default", "node":"172.31.20.193", "pod":"nginx-deployment-6d5f899847-hhn2p", "timestamp":"2025-02-13 15:37:49.162573614 +0000 UTC"}, Hostname:"172.31.20.193", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.192 [INFO][3616] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3616] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.255 [INFO][3616] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.20.193' Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.261 [INFO][3616] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.271 [INFO][3616] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.294 [INFO][3616] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.311 [INFO][3616] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.324 [INFO][3616] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.324 [INFO][3616] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.329 [INFO][3616] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4 Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.336 [INFO][3616] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.349 [INFO][3616] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.66/26] block=192.168.106.64/26 handle="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.349 [INFO][3616] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.66/26] handle="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" host="172.31.20.193" Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.349 [INFO][3616] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 15:37:49.392458 containerd[1992]: 2025-02-13 15:37:49.349 [INFO][3616] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.66/26] IPv6=[] ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" HandleID="k8s-pod-network.df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Workload="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.353 [INFO][3599] cni-plugin/k8s.go 386: Populated endpoint ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"a90b5295-846c-4b90-98b2-8b1ab6c5419a", ResourceVersion:"1049", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 37, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"", Pod:"nginx-deployment-6d5f899847-hhn2p", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.106.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali366bf208cdf", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.353 [INFO][3599] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.66/32] ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.353 [INFO][3599] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali366bf208cdf ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.362 [INFO][3599] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.364 [INFO][3599] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0", GenerateName:"nginx-deployment-6d5f899847-", Namespace:"default", SelfLink:"", UID:"a90b5295-846c-4b90-98b2-8b1ab6c5419a", ResourceVersion:"1049", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 37, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"6d5f899847", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4", Pod:"nginx-deployment-6d5f899847-hhn2p", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.106.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali366bf208cdf", MAC:"02:65:d2:18:46:21", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:37:49.394175 containerd[1992]: 2025-02-13 15:37:49.379 [INFO][3599] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4" Namespace="default" Pod="nginx-deployment-6d5f899847-hhn2p" WorkloadEndpoint="172.31.20.193-k8s-nginx--deployment--6d5f899847--hhn2p-eth0" Feb 13 15:37:49.467052 containerd[1992]: time="2025-02-13T15:37:49.467008610Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-kxrfd,Uid:9afcbeb8-f8e4-4f2f-9b35-6cef169285f0,Namespace:calico-system,Attempt:10,} returns sandbox id \"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5\"" Feb 13 15:37:49.471493 containerd[1992]: time="2025-02-13T15:37:49.471225828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Feb 13 15:37:49.478421 containerd[1992]: time="2025-02-13T15:37:49.478309565Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:37:49.478421 containerd[1992]: time="2025-02-13T15:37:49.478361374Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:37:49.478421 containerd[1992]: time="2025-02-13T15:37:49.478379330Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:49.478914 containerd[1992]: time="2025-02-13T15:37:49.478683272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:37:49.574106 containerd[1992]: time="2025-02-13T15:37:49.573877056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-hhn2p,Uid:a90b5295-846c-4b90-98b2-8b1ab6c5419a,Namespace:default,Attempt:7,} returns sandbox id \"df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4\"" Feb 13 15:37:50.223301 kernel: bpftool[3853]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Feb 13 15:37:50.317511 kubelet[2499]: E0213 15:37:50.317454 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:50.426338 systemd-networkd[1568]: cali53e2623664b: Gained IPv6LL Feb 13 15:37:50.566811 systemd-networkd[1568]: vxlan.calico: Link UP Feb 13 15:37:50.566820 systemd-networkd[1568]: vxlan.calico: Gained carrier Feb 13 15:37:50.752011 systemd-networkd[1568]: cali366bf208cdf: Gained IPv6LL Feb 13 15:37:51.239513 containerd[1992]: time="2025-02-13T15:37:51.239462826Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:51.241328 containerd[1992]: time="2025-02-13T15:37:51.241195820Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Feb 13 15:37:51.243279 containerd[1992]: time="2025-02-13T15:37:51.243243710Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:51.247311 containerd[1992]: time="2025-02-13T15:37:51.246588239Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:51.247311 containerd[1992]: time="2025-02-13T15:37:51.247179920Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 1.77591167s" Feb 13 15:37:51.247311 containerd[1992]: time="2025-02-13T15:37:51.247212180Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Feb 13 15:37:51.248507 containerd[1992]: time="2025-02-13T15:37:51.248484055Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Feb 13 15:37:51.249277 containerd[1992]: time="2025-02-13T15:37:51.249246022Z" level=info msg="CreateContainer within sandbox \"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Feb 13 15:37:51.285743 containerd[1992]: time="2025-02-13T15:37:51.285526656Z" level=info msg="CreateContainer within sandbox \"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"d946bb8a70dff8b2542391efc5e489b24a967062c4e4b1ab84f5969fdc7474d2\"" Feb 13 15:37:51.287814 containerd[1992]: time="2025-02-13T15:37:51.286336632Z" level=info msg="StartContainer for \"d946bb8a70dff8b2542391efc5e489b24a967062c4e4b1ab84f5969fdc7474d2\"" Feb 13 15:37:51.319558 kubelet[2499]: E0213 15:37:51.319524 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:51.331294 systemd[1]: run-containerd-runc-k8s.io-d946bb8a70dff8b2542391efc5e489b24a967062c4e4b1ab84f5969fdc7474d2-runc.M0gKIX.mount: Deactivated successfully. Feb 13 15:37:51.388525 containerd[1992]: time="2025-02-13T15:37:51.388482247Z" level=info msg="StartContainer for \"d946bb8a70dff8b2542391efc5e489b24a967062c4e4b1ab84f5969fdc7474d2\" returns successfully" Feb 13 15:37:51.642923 systemd-networkd[1568]: vxlan.calico: Gained IPv6LL Feb 13 15:37:52.321111 kubelet[2499]: E0213 15:37:52.321048 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:52.850864 kubelet[2499]: I0213 15:37:52.850784 2499 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 15:37:53.322118 kubelet[2499]: E0213 15:37:53.322077 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:54.120663 ntpd[1955]: Listen normally on 6 vxlan.calico 192.168.106.64:123 Feb 13 15:37:54.121325 ntpd[1955]: Listen normally on 7 cali53e2623664b [fe80::ecee:eeff:feee:eeee%3]:123 Feb 13 15:37:54.126082 ntpd[1955]: 13 Feb 15:37:54 ntpd[1955]: Listen normally on 6 vxlan.calico 192.168.106.64:123 Feb 13 15:37:54.126082 ntpd[1955]: 13 Feb 15:37:54 ntpd[1955]: Listen normally on 7 cali53e2623664b [fe80::ecee:eeff:feee:eeee%3]:123 Feb 13 15:37:54.126082 ntpd[1955]: 13 Feb 15:37:54 ntpd[1955]: Listen normally on 8 cali366bf208cdf [fe80::ecee:eeff:feee:eeee%4]:123 Feb 13 15:37:54.126082 ntpd[1955]: 13 Feb 15:37:54 ntpd[1955]: Listen normally on 9 vxlan.calico [fe80::6424:91ff:fe4c:b3ac%5]:123 Feb 13 15:37:54.121550 ntpd[1955]: Listen normally on 8 cali366bf208cdf [fe80::ecee:eeff:feee:eeee%4]:123 Feb 13 15:37:54.121887 ntpd[1955]: Listen normally on 9 vxlan.calico [fe80::6424:91ff:fe4c:b3ac%5]:123 Feb 13 15:37:54.323187 kubelet[2499]: E0213 15:37:54.323146 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:55.117544 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1648110339.mount: Deactivated successfully. Feb 13 15:37:55.323432 kubelet[2499]: E0213 15:37:55.323396 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:56.326086 kubelet[2499]: E0213 15:37:56.326009 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:56.628175 update_engine[1968]: I20250213 15:37:56.628110 1968 update_attempter.cc:509] Updating boot flags... Feb 13 15:37:56.738197 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (4036) Feb 13 15:37:57.296090 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (4027) Feb 13 15:37:57.326765 kubelet[2499]: E0213 15:37:57.326727 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:57.649717 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (4027) Feb 13 15:37:57.715016 containerd[1992]: time="2025-02-13T15:37:57.714957088Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:57.717647 containerd[1992]: time="2025-02-13T15:37:57.717582532Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=73054493" Feb 13 15:37:57.719243 containerd[1992]: time="2025-02-13T15:37:57.719178444Z" level=info msg="ImageCreate event name:\"sha256:fe94eb5f0c9c8d0ca277aa8cd5940f1faf5970175bf373932babc578545deda8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:57.739496 containerd[1992]: time="2025-02-13T15:37:57.739424613Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:d9bc3da999da9f147f1277c7b18292486847e8f39f95fcf81d914d0c22815faf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:57.795849 containerd[1992]: time="2025-02-13T15:37:57.792898742Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fe94eb5f0c9c8d0ca277aa8cd5940f1faf5970175bf373932babc578545deda8\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:d9bc3da999da9f147f1277c7b18292486847e8f39f95fcf81d914d0c22815faf\", size \"73054371\" in 6.544237124s" Feb 13 15:37:57.795849 containerd[1992]: time="2025-02-13T15:37:57.792953565Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fe94eb5f0c9c8d0ca277aa8cd5940f1faf5970175bf373932babc578545deda8\"" Feb 13 15:37:57.799043 containerd[1992]: time="2025-02-13T15:37:57.796955878Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Feb 13 15:37:57.862926 containerd[1992]: time="2025-02-13T15:37:57.862887397Z" level=info msg="CreateContainer within sandbox \"df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Feb 13 15:37:57.886138 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2902813311.mount: Deactivated successfully. Feb 13 15:37:57.891094 containerd[1992]: time="2025-02-13T15:37:57.890811878Z" level=info msg="CreateContainer within sandbox \"df6c4da84adf3b0e22970682d7a3a8e9b12efe59638f4b4da552f41daaec30e4\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"4d59736865cb87de71b7ab95cd1c3704634782efda054d26f2cf510b869b486b\"" Feb 13 15:37:57.891873 containerd[1992]: time="2025-02-13T15:37:57.891833307Z" level=info msg="StartContainer for \"4d59736865cb87de71b7ab95cd1c3704634782efda054d26f2cf510b869b486b\"" Feb 13 15:37:57.964587 containerd[1992]: time="2025-02-13T15:37:57.964114948Z" level=info msg="StartContainer for \"4d59736865cb87de71b7ab95cd1c3704634782efda054d26f2cf510b869b486b\" returns successfully" Feb 13 15:37:58.327294 kubelet[2499]: E0213 15:37:58.326957 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:59.330154 kubelet[2499]: E0213 15:37:59.327456 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:37:59.704355 containerd[1992]: time="2025-02-13T15:37:59.704308537Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:59.707697 containerd[1992]: time="2025-02-13T15:37:59.707529843Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Feb 13 15:37:59.710107 containerd[1992]: time="2025-02-13T15:37:59.709397259Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:59.718196 containerd[1992]: time="2025-02-13T15:37:59.717566179Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:37:59.720511 containerd[1992]: time="2025-02-13T15:37:59.720396714Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 1.923387387s" Feb 13 15:37:59.720654 containerd[1992]: time="2025-02-13T15:37:59.720517663Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Feb 13 15:37:59.725480 containerd[1992]: time="2025-02-13T15:37:59.725438670Z" level=info msg="CreateContainer within sandbox \"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Feb 13 15:37:59.779545 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2330136082.mount: Deactivated successfully. Feb 13 15:37:59.823951 containerd[1992]: time="2025-02-13T15:37:59.823898298Z" level=info msg="CreateContainer within sandbox \"394d232719bd3862310643f2c7c8cc6ab845e3efc1413da68d35d2eec82ca0e5\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"9fdd9000e71e26889d03ce9313212aee4fbb666493f4296cb08bcc17f2c180cc\"" Feb 13 15:37:59.824686 containerd[1992]: time="2025-02-13T15:37:59.824649737Z" level=info msg="StartContainer for \"9fdd9000e71e26889d03ce9313212aee4fbb666493f4296cb08bcc17f2c180cc\"" Feb 13 15:37:59.881957 systemd[1]: run-containerd-runc-k8s.io-9fdd9000e71e26889d03ce9313212aee4fbb666493f4296cb08bcc17f2c180cc-runc.jsaeYv.mount: Deactivated successfully. Feb 13 15:37:59.944327 containerd[1992]: time="2025-02-13T15:37:59.942610029Z" level=info msg="StartContainer for \"9fdd9000e71e26889d03ce9313212aee4fbb666493f4296cb08bcc17f2c180cc\" returns successfully" Feb 13 15:38:00.175760 kubelet[2499]: I0213 15:38:00.175593 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/csi-node-driver-kxrfd" podStartSLOduration=25.923156011 podStartE2EDuration="36.175532939s" podCreationTimestamp="2025-02-13 15:37:24 +0000 UTC" firstStartedPulling="2025-02-13 15:37:49.470949468 +0000 UTC m=+25.514594958" lastFinishedPulling="2025-02-13 15:37:59.723326402 +0000 UTC m=+35.766971886" observedRunningTime="2025-02-13 15:38:00.172285752 +0000 UTC m=+36.215931255" watchObservedRunningTime="2025-02-13 15:38:00.175532939 +0000 UTC m=+36.219178445" Feb 13 15:38:00.176010 kubelet[2499]: I0213 15:38:00.175908 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-hhn2p" podStartSLOduration=10.96310965 podStartE2EDuration="19.175871578s" podCreationTimestamp="2025-02-13 15:37:41 +0000 UTC" firstStartedPulling="2025-02-13 15:37:49.580603631 +0000 UTC m=+25.624249125" lastFinishedPulling="2025-02-13 15:37:57.793365556 +0000 UTC m=+33.837011053" observedRunningTime="2025-02-13 15:37:58.052753029 +0000 UTC m=+34.096398538" watchObservedRunningTime="2025-02-13 15:38:00.175871578 +0000 UTC m=+36.219517085" Feb 13 15:38:00.328218 kubelet[2499]: E0213 15:38:00.328176 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:00.473032 kubelet[2499]: I0213 15:38:00.472897 2499 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Feb 13 15:38:00.474772 kubelet[2499]: I0213 15:38:00.474739 2499 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Feb 13 15:38:01.328620 kubelet[2499]: E0213 15:38:01.328563 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:02.329263 kubelet[2499]: E0213 15:38:02.329206 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:03.330148 kubelet[2499]: E0213 15:38:03.329983 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:04.283203 kubelet[2499]: E0213 15:38:04.283148 2499 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:04.330899 kubelet[2499]: E0213 15:38:04.330854 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:05.331767 kubelet[2499]: E0213 15:38:05.331712 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:06.332714 kubelet[2499]: E0213 15:38:06.332652 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:07.333317 kubelet[2499]: E0213 15:38:07.333264 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:07.655905 kubelet[2499]: I0213 15:38:07.655789 2499 topology_manager.go:215] "Topology Admit Handler" podUID="e9f2e108-61ef-48db-8f36-809ba30e8324" podNamespace="default" podName="nfs-server-provisioner-0" Feb 13 15:38:07.781227 kubelet[2499]: I0213 15:38:07.781163 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/e9f2e108-61ef-48db-8f36-809ba30e8324-data\") pod \"nfs-server-provisioner-0\" (UID: \"e9f2e108-61ef-48db-8f36-809ba30e8324\") " pod="default/nfs-server-provisioner-0" Feb 13 15:38:07.800216 kubelet[2499]: I0213 15:38:07.800164 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hhwj\" (UniqueName: \"kubernetes.io/projected/e9f2e108-61ef-48db-8f36-809ba30e8324-kube-api-access-5hhwj\") pod \"nfs-server-provisioner-0\" (UID: \"e9f2e108-61ef-48db-8f36-809ba30e8324\") " pod="default/nfs-server-provisioner-0" Feb 13 15:38:07.966391 containerd[1992]: time="2025-02-13T15:38:07.964351167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:e9f2e108-61ef-48db-8f36-809ba30e8324,Namespace:default,Attempt:0,}" Feb 13 15:38:08.231466 systemd-networkd[1568]: cali60e51b789ff: Link UP Feb 13 15:38:08.231694 systemd-networkd[1568]: cali60e51b789ff: Gained carrier Feb 13 15:38:08.246844 (udev-worker)[4436]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.076 [INFO][4418] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.20.193-k8s-nfs--server--provisioner--0-eth0 nfs-server-provisioner- default e9f2e108-61ef-48db-8f36-809ba30e8324 1216 0 2025-02-13 15:38:07 +0000 UTC map[app:nfs-server-provisioner apps.kubernetes.io/pod-index:0 chart:nfs-server-provisioner-1.8.0 controller-revision-hash:nfs-server-provisioner-d5cbb7f57 heritage:Helm projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:nfs-server-provisioner release:nfs-server-provisioner statefulset.kubernetes.io/pod-name:nfs-server-provisioner-0] map[] [] [] []} {k8s 172.31.20.193 nfs-server-provisioner-0 eth0 nfs-server-provisioner [] [] [kns.default ksa.default.nfs-server-provisioner] cali60e51b789ff [{nfs TCP 2049 0 } {nfs-udp UDP 2049 0 } {nlockmgr TCP 32803 0 } {nlockmgr-udp UDP 32803 0 } {mountd TCP 20048 0 } {mountd-udp UDP 20048 0 } {rquotad TCP 875 0 } {rquotad-udp UDP 875 0 } {rpcbind TCP 111 0 } {rpcbind-udp UDP 111 0 } {statd TCP 662 0 } {statd-udp UDP 662 0 }] []}} ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.076 [INFO][4418] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.129 [INFO][4429] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" HandleID="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Workload="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.158 [INFO][4429] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" HandleID="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Workload="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000291520), Attrs:map[string]string{"namespace":"default", "node":"172.31.20.193", "pod":"nfs-server-provisioner-0", "timestamp":"2025-02-13 15:38:08.129781385 +0000 UTC"}, Hostname:"172.31.20.193", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.158 [INFO][4429] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.158 [INFO][4429] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.158 [INFO][4429] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.20.193' Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.161 [INFO][4429] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.173 [INFO][4429] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.182 [INFO][4429] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.187 [INFO][4429] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.191 [INFO][4429] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.191 [INFO][4429] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.199 [INFO][4429] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.206 [INFO][4429] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.221 [INFO][4429] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.67/26] block=192.168.106.64/26 handle="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.221 [INFO][4429] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.67/26] handle="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" host="172.31.20.193" Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.221 [INFO][4429] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 15:38:08.271844 containerd[1992]: 2025-02-13 15:38:08.221 [INFO][4429] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.67/26] IPv6=[] ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" HandleID="k8s-pod-network.0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Workload="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.273154 containerd[1992]: 2025-02-13 15:38:08.223 [INFO][4418] cni-plugin/k8s.go 386: Populated endpoint ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"e9f2e108-61ef-48db-8f36-809ba30e8324", ResourceVersion:"1216", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 38, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.106.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:38:08.273154 containerd[1992]: 2025-02-13 15:38:08.224 [INFO][4418] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.67/32] ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.273154 containerd[1992]: 2025-02-13 15:38:08.224 [INFO][4418] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali60e51b789ff ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.273154 containerd[1992]: 2025-02-13 15:38:08.233 [INFO][4418] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.273520 containerd[1992]: 2025-02-13 15:38:08.233 [INFO][4418] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"e9f2e108-61ef-48db-8f36-809ba30e8324", ResourceVersion:"1216", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 38, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.106.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"de:71:d4:ac:b3:21", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:38:08.273520 containerd[1992]: 2025-02-13 15:38:08.259 [INFO][4418] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.20.193-k8s-nfs--server--provisioner--0-eth0" Feb 13 15:38:08.322393 containerd[1992]: time="2025-02-13T15:38:08.321339338Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:38:08.322393 containerd[1992]: time="2025-02-13T15:38:08.321404365Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:38:08.322393 containerd[1992]: time="2025-02-13T15:38:08.321421130Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:38:08.322393 containerd[1992]: time="2025-02-13T15:38:08.321516881Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:38:08.335685 kubelet[2499]: E0213 15:38:08.334118 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:08.447266 containerd[1992]: time="2025-02-13T15:38:08.447225006Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:e9f2e108-61ef-48db-8f36-809ba30e8324,Namespace:default,Attempt:0,} returns sandbox id \"0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e\"" Feb 13 15:38:08.449551 containerd[1992]: time="2025-02-13T15:38:08.449188914Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Feb 13 15:38:09.336366 kubelet[2499]: E0213 15:38:09.336320 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:09.818337 systemd-networkd[1568]: cali60e51b789ff: Gained IPv6LL Feb 13 15:38:10.338405 kubelet[2499]: E0213 15:38:10.338359 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:11.339173 kubelet[2499]: E0213 15:38:11.339142 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:11.619585 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1990150670.mount: Deactivated successfully. Feb 13 15:38:12.114386 ntpd[1955]: Listen normally on 10 cali60e51b789ff [fe80::ecee:eeff:feee:eeee%8]:123 Feb 13 15:38:12.118661 ntpd[1955]: 13 Feb 15:38:12 ntpd[1955]: Listen normally on 10 cali60e51b789ff [fe80::ecee:eeff:feee:eeee%8]:123 Feb 13 15:38:12.342151 kubelet[2499]: E0213 15:38:12.340801 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:13.341863 kubelet[2499]: E0213 15:38:13.341825 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:14.342331 kubelet[2499]: E0213 15:38:14.342292 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:14.820376 containerd[1992]: time="2025-02-13T15:38:14.820297263Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:38:14.822325 containerd[1992]: time="2025-02-13T15:38:14.822092061Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=91039406" Feb 13 15:38:14.825107 containerd[1992]: time="2025-02-13T15:38:14.824576078Z" level=info msg="ImageCreate event name:\"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:38:14.828778 containerd[1992]: time="2025-02-13T15:38:14.828734232Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:38:14.829944 containerd[1992]: time="2025-02-13T15:38:14.829902729Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"91036984\" in 6.380676373s" Feb 13 15:38:14.830111 containerd[1992]: time="2025-02-13T15:38:14.830091253Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Feb 13 15:38:14.832288 containerd[1992]: time="2025-02-13T15:38:14.832256086Z" level=info msg="CreateContainer within sandbox \"0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Feb 13 15:38:14.869100 containerd[1992]: time="2025-02-13T15:38:14.868986502Z" level=info msg="CreateContainer within sandbox \"0a0c400087b5cdf65d412129627b5b1b10067ada3d721bbd3c1e046ee122755e\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"424206e9a01155a662cf1a21746ad4f0ad88f064f33d441194a1c64e20ce9223\"" Feb 13 15:38:14.872234 containerd[1992]: time="2025-02-13T15:38:14.870916617Z" level=info msg="StartContainer for \"424206e9a01155a662cf1a21746ad4f0ad88f064f33d441194a1c64e20ce9223\"" Feb 13 15:38:14.922327 systemd[1]: run-containerd-runc-k8s.io-424206e9a01155a662cf1a21746ad4f0ad88f064f33d441194a1c64e20ce9223-runc.3LQUXj.mount: Deactivated successfully. Feb 13 15:38:14.960798 containerd[1992]: time="2025-02-13T15:38:14.960664392Z" level=info msg="StartContainer for \"424206e9a01155a662cf1a21746ad4f0ad88f064f33d441194a1c64e20ce9223\" returns successfully" Feb 13 15:38:15.344104 kubelet[2499]: E0213 15:38:15.344044 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:15.382044 kubelet[2499]: I0213 15:38:15.382008 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=2.000349784 podStartE2EDuration="8.381963994s" podCreationTimestamp="2025-02-13 15:38:07 +0000 UTC" firstStartedPulling="2025-02-13 15:38:08.448803904 +0000 UTC m=+44.492449396" lastFinishedPulling="2025-02-13 15:38:14.83041811 +0000 UTC m=+50.874063606" observedRunningTime="2025-02-13 15:38:15.381689909 +0000 UTC m=+51.425335416" watchObservedRunningTime="2025-02-13 15:38:15.381963994 +0000 UTC m=+51.425609499" Feb 13 15:38:16.344460 kubelet[2499]: E0213 15:38:16.344401 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:17.345080 kubelet[2499]: E0213 15:38:17.345015 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:18.345959 kubelet[2499]: E0213 15:38:18.345907 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:19.346223 kubelet[2499]: E0213 15:38:19.346140 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:20.346816 kubelet[2499]: E0213 15:38:20.346759 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:21.347735 kubelet[2499]: E0213 15:38:21.347678 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:22.348571 kubelet[2499]: E0213 15:38:22.348516 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:23.349290 kubelet[2499]: E0213 15:38:23.349238 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:24.282619 kubelet[2499]: E0213 15:38:24.282567 2499 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:24.350414 kubelet[2499]: E0213 15:38:24.350367 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:24.352490 containerd[1992]: time="2025-02-13T15:38:24.352449391Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:38:24.354371 containerd[1992]: time="2025-02-13T15:38:24.352578320Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:38:24.354371 containerd[1992]: time="2025-02-13T15:38:24.352595100Z" level=info msg="StopPodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:38:24.369712 containerd[1992]: time="2025-02-13T15:38:24.369660475Z" level=info msg="RemovePodSandbox for \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:38:24.376406 containerd[1992]: time="2025-02-13T15:38:24.376354851Z" level=info msg="Forcibly stopping sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\"" Feb 13 15:38:24.376559 containerd[1992]: time="2025-02-13T15:38:24.376498222Z" level=info msg="TearDown network for sandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" successfully" Feb 13 15:38:24.396318 containerd[1992]: time="2025-02-13T15:38:24.395982827Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.396318 containerd[1992]: time="2025-02-13T15:38:24.396206617Z" level=info msg="RemovePodSandbox \"394abbb0c3efdb1cc636a7a2b721b009b0a68d0ebd7b15689b9cf65cf6438d81\" returns successfully" Feb 13 15:38:24.396689 containerd[1992]: time="2025-02-13T15:38:24.396658013Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:38:24.396793 containerd[1992]: time="2025-02-13T15:38:24.396767732Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:38:24.396853 containerd[1992]: time="2025-02-13T15:38:24.396788892Z" level=info msg="StopPodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:38:24.399054 containerd[1992]: time="2025-02-13T15:38:24.397252991Z" level=info msg="RemovePodSandbox for \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:38:24.399054 containerd[1992]: time="2025-02-13T15:38:24.397283948Z" level=info msg="Forcibly stopping sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\"" Feb 13 15:38:24.399054 containerd[1992]: time="2025-02-13T15:38:24.397364846Z" level=info msg="TearDown network for sandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" successfully" Feb 13 15:38:24.405962 containerd[1992]: time="2025-02-13T15:38:24.405842771Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.406266 containerd[1992]: time="2025-02-13T15:38:24.406241521Z" level=info msg="RemovePodSandbox \"2d820ac90165ddf5242a9b15e553babbf89c8d9dd5f42c83bbab1bb0891fbc85\" returns successfully" Feb 13 15:38:24.406854 containerd[1992]: time="2025-02-13T15:38:24.406826701Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:38:24.407150 containerd[1992]: time="2025-02-13T15:38:24.407129593Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:38:24.407316 containerd[1992]: time="2025-02-13T15:38:24.407298024Z" level=info msg="StopPodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:38:24.407874 containerd[1992]: time="2025-02-13T15:38:24.407837596Z" level=info msg="RemovePodSandbox for \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:38:24.407947 containerd[1992]: time="2025-02-13T15:38:24.407872798Z" level=info msg="Forcibly stopping sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\"" Feb 13 15:38:24.407999 containerd[1992]: time="2025-02-13T15:38:24.407954604Z" level=info msg="TearDown network for sandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" successfully" Feb 13 15:38:24.413621 containerd[1992]: time="2025-02-13T15:38:24.413544302Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.414313 containerd[1992]: time="2025-02-13T15:38:24.413640922Z" level=info msg="RemovePodSandbox \"dae3fa116ca55fa2b49013fd7741d6b6505a5a781967828dac70650a74486aab\" returns successfully" Feb 13 15:38:24.414693 containerd[1992]: time="2025-02-13T15:38:24.414665063Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:38:24.414802 containerd[1992]: time="2025-02-13T15:38:24.414778338Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:38:24.414850 containerd[1992]: time="2025-02-13T15:38:24.414799705Z" level=info msg="StopPodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:38:24.415687 containerd[1992]: time="2025-02-13T15:38:24.415619704Z" level=info msg="RemovePodSandbox for \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:38:24.415687 containerd[1992]: time="2025-02-13T15:38:24.415652603Z" level=info msg="Forcibly stopping sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\"" Feb 13 15:38:24.415879 containerd[1992]: time="2025-02-13T15:38:24.415808699Z" level=info msg="TearDown network for sandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" successfully" Feb 13 15:38:24.424292 containerd[1992]: time="2025-02-13T15:38:24.424222974Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.424292 containerd[1992]: time="2025-02-13T15:38:24.424293446Z" level=info msg="RemovePodSandbox \"6c7926054ff4c70ca0c49ff102a9d18e12b18ef2d2d3f527cd6bd223866b3689\" returns successfully" Feb 13 15:38:24.425447 containerd[1992]: time="2025-02-13T15:38:24.425407160Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:38:24.425689 containerd[1992]: time="2025-02-13T15:38:24.425643588Z" level=info msg="TearDown network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" successfully" Feb 13 15:38:24.425689 containerd[1992]: time="2025-02-13T15:38:24.425666343Z" level=info msg="StopPodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" returns successfully" Feb 13 15:38:24.426704 containerd[1992]: time="2025-02-13T15:38:24.426176397Z" level=info msg="RemovePodSandbox for \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:38:24.426864 containerd[1992]: time="2025-02-13T15:38:24.426712562Z" level=info msg="Forcibly stopping sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\"" Feb 13 15:38:24.427872 containerd[1992]: time="2025-02-13T15:38:24.426884023Z" level=info msg="TearDown network for sandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" successfully" Feb 13 15:38:24.434757 containerd[1992]: time="2025-02-13T15:38:24.434598960Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.434757 containerd[1992]: time="2025-02-13T15:38:24.434660855Z" level=info msg="RemovePodSandbox \"3f18c7ae28b6b1b8708a697a222dde8639d8671c7ae6c2bd77a94774d07947c6\" returns successfully" Feb 13 15:38:24.435331 containerd[1992]: time="2025-02-13T15:38:24.435304006Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" Feb 13 15:38:24.435601 containerd[1992]: time="2025-02-13T15:38:24.435576034Z" level=info msg="TearDown network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" successfully" Feb 13 15:38:24.435601 containerd[1992]: time="2025-02-13T15:38:24.435596148Z" level=info msg="StopPodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" returns successfully" Feb 13 15:38:24.436299 containerd[1992]: time="2025-02-13T15:38:24.436275043Z" level=info msg="RemovePodSandbox for \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" Feb 13 15:38:24.436411 containerd[1992]: time="2025-02-13T15:38:24.436302782Z" level=info msg="Forcibly stopping sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\"" Feb 13 15:38:24.436482 containerd[1992]: time="2025-02-13T15:38:24.436426699Z" level=info msg="TearDown network for sandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" successfully" Feb 13 15:38:24.441450 containerd[1992]: time="2025-02-13T15:38:24.441408794Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.441660 containerd[1992]: time="2025-02-13T15:38:24.441464529Z" level=info msg="RemovePodSandbox \"af2406d69c9780ca08f6a9e9b1fbb29d08874bb89bd68c5a2671a9942ec47ea7\" returns successfully" Feb 13 15:38:24.442991 containerd[1992]: time="2025-02-13T15:38:24.442453720Z" level=info msg="StopPodSandbox for \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\"" Feb 13 15:38:24.442991 containerd[1992]: time="2025-02-13T15:38:24.442876400Z" level=info msg="TearDown network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" successfully" Feb 13 15:38:24.442991 containerd[1992]: time="2025-02-13T15:38:24.442894460Z" level=info msg="StopPodSandbox for \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" returns successfully" Feb 13 15:38:24.443372 containerd[1992]: time="2025-02-13T15:38:24.443347051Z" level=info msg="RemovePodSandbox for \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\"" Feb 13 15:38:24.443454 containerd[1992]: time="2025-02-13T15:38:24.443375637Z" level=info msg="Forcibly stopping sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\"" Feb 13 15:38:24.443515 containerd[1992]: time="2025-02-13T15:38:24.443458921Z" level=info msg="TearDown network for sandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" successfully" Feb 13 15:38:24.448469 containerd[1992]: time="2025-02-13T15:38:24.448420362Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.448592 containerd[1992]: time="2025-02-13T15:38:24.448478844Z" level=info msg="RemovePodSandbox \"b5feae9d9354ab01a13ada8cf443a8298e1a69ca9965ed7bc9a674f9ef5fe476\" returns successfully" Feb 13 15:38:24.448988 containerd[1992]: time="2025-02-13T15:38:24.448962212Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:38:24.449237 containerd[1992]: time="2025-02-13T15:38:24.449161049Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:38:24.449237 containerd[1992]: time="2025-02-13T15:38:24.449180179Z" level=info msg="StopPodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:38:24.450061 containerd[1992]: time="2025-02-13T15:38:24.449907021Z" level=info msg="RemovePodSandbox for \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:38:24.450609 containerd[1992]: time="2025-02-13T15:38:24.450115021Z" level=info msg="Forcibly stopping sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\"" Feb 13 15:38:24.450718 containerd[1992]: time="2025-02-13T15:38:24.450662631Z" level=info msg="TearDown network for sandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" successfully" Feb 13 15:38:24.455949 containerd[1992]: time="2025-02-13T15:38:24.455901923Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.456301 containerd[1992]: time="2025-02-13T15:38:24.455956975Z" level=info msg="RemovePodSandbox \"af24380772e4b2d57c5aaf0aed4dc5b76179df9aa6ff0c72f2763b418e86effe\" returns successfully" Feb 13 15:38:24.456588 containerd[1992]: time="2025-02-13T15:38:24.456551298Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:38:24.456696 containerd[1992]: time="2025-02-13T15:38:24.456669009Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:38:24.456750 containerd[1992]: time="2025-02-13T15:38:24.456693749Z" level=info msg="StopPodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:38:24.457078 containerd[1992]: time="2025-02-13T15:38:24.457038335Z" level=info msg="RemovePodSandbox for \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:38:24.457145 containerd[1992]: time="2025-02-13T15:38:24.457094445Z" level=info msg="Forcibly stopping sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\"" Feb 13 15:38:24.457225 containerd[1992]: time="2025-02-13T15:38:24.457176123Z" level=info msg="TearDown network for sandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" successfully" Feb 13 15:38:24.463502 containerd[1992]: time="2025-02-13T15:38:24.463450462Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.463851 containerd[1992]: time="2025-02-13T15:38:24.463516719Z" level=info msg="RemovePodSandbox \"3cabb71ca8358aae6c2dadf2c6e023c92165e1113cc0183f232304d8808c5800\" returns successfully" Feb 13 15:38:24.464102 containerd[1992]: time="2025-02-13T15:38:24.464053483Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:38:24.464215 containerd[1992]: time="2025-02-13T15:38:24.464190000Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:38:24.464269 containerd[1992]: time="2025-02-13T15:38:24.464212685Z" level=info msg="StopPodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:38:24.464710 containerd[1992]: time="2025-02-13T15:38:24.464609526Z" level=info msg="RemovePodSandbox for \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:38:24.464710 containerd[1992]: time="2025-02-13T15:38:24.464642310Z" level=info msg="Forcibly stopping sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\"" Feb 13 15:38:24.465822 containerd[1992]: time="2025-02-13T15:38:24.464921232Z" level=info msg="TearDown network for sandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" successfully" Feb 13 15:38:24.475614 containerd[1992]: time="2025-02-13T15:38:24.475557798Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.475789 containerd[1992]: time="2025-02-13T15:38:24.475761680Z" level=info msg="RemovePodSandbox \"e8519115339707c4764df05621d4e7af83219244d29aeb50a1843cbd38c31f71\" returns successfully" Feb 13 15:38:24.476405 containerd[1992]: time="2025-02-13T15:38:24.476369076Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:38:24.476510 containerd[1992]: time="2025-02-13T15:38:24.476491811Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:38:24.476556 containerd[1992]: time="2025-02-13T15:38:24.476511049Z" level=info msg="StopPodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:38:24.478117 containerd[1992]: time="2025-02-13T15:38:24.476825874Z" level=info msg="RemovePodSandbox for \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:38:24.478117 containerd[1992]: time="2025-02-13T15:38:24.476856304Z" level=info msg="Forcibly stopping sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\"" Feb 13 15:38:24.478117 containerd[1992]: time="2025-02-13T15:38:24.477774694Z" level=info msg="TearDown network for sandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" successfully" Feb 13 15:38:24.494793 containerd[1992]: time="2025-02-13T15:38:24.493566180Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.494793 containerd[1992]: time="2025-02-13T15:38:24.493860537Z" level=info msg="RemovePodSandbox \"ccbc142eb3b06998eb8b6e23a2102e4a34085a3fda40e017d901a012c9293bfe\" returns successfully" Feb 13 15:38:24.497378 containerd[1992]: time="2025-02-13T15:38:24.497130679Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:38:24.497378 containerd[1992]: time="2025-02-13T15:38:24.497299148Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:38:24.497378 containerd[1992]: time="2025-02-13T15:38:24.497326897Z" level=info msg="StopPodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:38:24.497939 containerd[1992]: time="2025-02-13T15:38:24.497911733Z" level=info msg="RemovePodSandbox for \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:38:24.498105 containerd[1992]: time="2025-02-13T15:38:24.497941417Z" level=info msg="Forcibly stopping sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\"" Feb 13 15:38:24.498478 containerd[1992]: time="2025-02-13T15:38:24.498115440Z" level=info msg="TearDown network for sandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" successfully" Feb 13 15:38:24.509127 containerd[1992]: time="2025-02-13T15:38:24.509058532Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.509294 containerd[1992]: time="2025-02-13T15:38:24.509144224Z" level=info msg="RemovePodSandbox \"c57f9858b2a9fd3a0115fe4b6c60310cf9e9929b3363a148e9fe4a0626e5d11d\" returns successfully" Feb 13 15:38:24.509800 containerd[1992]: time="2025-02-13T15:38:24.509718714Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:38:24.511021 containerd[1992]: time="2025-02-13T15:38:24.509838819Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:38:24.511118 containerd[1992]: time="2025-02-13T15:38:24.511013293Z" level=info msg="StopPodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:38:24.512107 containerd[1992]: time="2025-02-13T15:38:24.511648919Z" level=info msg="RemovePodSandbox for \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:38:24.512107 containerd[1992]: time="2025-02-13T15:38:24.511680001Z" level=info msg="Forcibly stopping sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\"" Feb 13 15:38:24.512107 containerd[1992]: time="2025-02-13T15:38:24.511768432Z" level=info msg="TearDown network for sandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" successfully" Feb 13 15:38:24.519707 containerd[1992]: time="2025-02-13T15:38:24.519646202Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.519931 containerd[1992]: time="2025-02-13T15:38:24.519713753Z" level=info msg="RemovePodSandbox \"4cb47b4405e7a98577cba4818224d40503c75d6528ce8e130f4820889a72d014\" returns successfully" Feb 13 15:38:24.520267 containerd[1992]: time="2025-02-13T15:38:24.520222835Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:38:24.520538 containerd[1992]: time="2025-02-13T15:38:24.520512361Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:38:24.520612 containerd[1992]: time="2025-02-13T15:38:24.520540304Z" level=info msg="StopPodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:38:24.521030 containerd[1992]: time="2025-02-13T15:38:24.520999763Z" level=info msg="RemovePodSandbox for \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:38:24.521124 containerd[1992]: time="2025-02-13T15:38:24.521032657Z" level=info msg="Forcibly stopping sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\"" Feb 13 15:38:24.521183 containerd[1992]: time="2025-02-13T15:38:24.521134658Z" level=info msg="TearDown network for sandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" successfully" Feb 13 15:38:24.527818 containerd[1992]: time="2025-02-13T15:38:24.527643365Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.527818 containerd[1992]: time="2025-02-13T15:38:24.527721462Z" level=info msg="RemovePodSandbox \"3f36d6930183f4a62f6499c14fb221087083edbf14370262a8b828acfafdbc93\" returns successfully" Feb 13 15:38:24.528194 containerd[1992]: time="2025-02-13T15:38:24.528165445Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:38:24.528301 containerd[1992]: time="2025-02-13T15:38:24.528275965Z" level=info msg="TearDown network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" successfully" Feb 13 15:38:24.528301 containerd[1992]: time="2025-02-13T15:38:24.528295859Z" level=info msg="StopPodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" returns successfully" Feb 13 15:38:24.528804 containerd[1992]: time="2025-02-13T15:38:24.528776719Z" level=info msg="RemovePodSandbox for \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:38:24.528882 containerd[1992]: time="2025-02-13T15:38:24.528807327Z" level=info msg="Forcibly stopping sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\"" Feb 13 15:38:24.528956 containerd[1992]: time="2025-02-13T15:38:24.528932691Z" level=info msg="TearDown network for sandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" successfully" Feb 13 15:38:24.533848 containerd[1992]: time="2025-02-13T15:38:24.533728329Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.533848 containerd[1992]: time="2025-02-13T15:38:24.533788775Z" level=info msg="RemovePodSandbox \"88beb5387ffaf68c22bb60fb8cc63012eacd57060cefbdd1447eda2fba37e3bf\" returns successfully" Feb 13 15:38:24.537231 containerd[1992]: time="2025-02-13T15:38:24.536591391Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" Feb 13 15:38:24.537231 containerd[1992]: time="2025-02-13T15:38:24.536948531Z" level=info msg="TearDown network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" successfully" Feb 13 15:38:24.537231 containerd[1992]: time="2025-02-13T15:38:24.536970639Z" level=info msg="StopPodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" returns successfully" Feb 13 15:38:24.538372 containerd[1992]: time="2025-02-13T15:38:24.537402450Z" level=info msg="RemovePodSandbox for \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" Feb 13 15:38:24.538372 containerd[1992]: time="2025-02-13T15:38:24.537430645Z" level=info msg="Forcibly stopping sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\"" Feb 13 15:38:24.538372 containerd[1992]: time="2025-02-13T15:38:24.537522504Z" level=info msg="TearDown network for sandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" successfully" Feb 13 15:38:24.543597 containerd[1992]: time="2025-02-13T15:38:24.543539752Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.543744 containerd[1992]: time="2025-02-13T15:38:24.543626979Z" level=info msg="RemovePodSandbox \"c85270c60411fbffad010968559b57961356f4de0dd3eb9f0efd1f0fa046ba47\" returns successfully" Feb 13 15:38:24.544123 containerd[1992]: time="2025-02-13T15:38:24.544089590Z" level=info msg="StopPodSandbox for \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\"" Feb 13 15:38:24.544232 containerd[1992]: time="2025-02-13T15:38:24.544206999Z" level=info msg="TearDown network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" successfully" Feb 13 15:38:24.544278 containerd[1992]: time="2025-02-13T15:38:24.544230699Z" level=info msg="StopPodSandbox for \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" returns successfully" Feb 13 15:38:24.544751 containerd[1992]: time="2025-02-13T15:38:24.544618708Z" level=info msg="RemovePodSandbox for \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\"" Feb 13 15:38:24.544751 containerd[1992]: time="2025-02-13T15:38:24.544648805Z" level=info msg="Forcibly stopping sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\"" Feb 13 15:38:24.544879 containerd[1992]: time="2025-02-13T15:38:24.544790222Z" level=info msg="TearDown network for sandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" successfully" Feb 13 15:38:24.549726 containerd[1992]: time="2025-02-13T15:38:24.549669261Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 15:38:24.549894 containerd[1992]: time="2025-02-13T15:38:24.549734075Z" level=info msg="RemovePodSandbox \"9e6f85b3235d601b712a43b4e5b0db4530abfe29014c80aba3e16e4e997988d8\" returns successfully" Feb 13 15:38:25.351935 kubelet[2499]: E0213 15:38:25.351800 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:26.353054 kubelet[2499]: E0213 15:38:26.353012 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:27.354145 kubelet[2499]: E0213 15:38:27.354003 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:28.355235 kubelet[2499]: E0213 15:38:28.355190 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:29.355389 kubelet[2499]: E0213 15:38:29.355332 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:30.356418 kubelet[2499]: E0213 15:38:30.356315 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:31.356905 kubelet[2499]: E0213 15:38:31.356845 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:32.357496 kubelet[2499]: E0213 15:38:32.357435 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:33.358823 kubelet[2499]: E0213 15:38:33.358698 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:34.359543 kubelet[2499]: E0213 15:38:34.359491 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:35.360736 kubelet[2499]: E0213 15:38:35.360679 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:36.361361 kubelet[2499]: E0213 15:38:36.361247 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:37.362314 kubelet[2499]: E0213 15:38:37.362254 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:38.363105 kubelet[2499]: E0213 15:38:38.363047 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:39.364043 kubelet[2499]: E0213 15:38:39.363985 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:39.417498 kubelet[2499]: I0213 15:38:39.417459 2499 topology_manager.go:215] "Topology Admit Handler" podUID="a3cfabd3-e9ad-461e-b47b-eb3e70a258b1" podNamespace="default" podName="test-pod-1" Feb 13 15:38:39.535431 kubelet[2499]: I0213 15:38:39.535378 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-426m7\" (UniqueName: \"kubernetes.io/projected/a3cfabd3-e9ad-461e-b47b-eb3e70a258b1-kube-api-access-426m7\") pod \"test-pod-1\" (UID: \"a3cfabd3-e9ad-461e-b47b-eb3e70a258b1\") " pod="default/test-pod-1" Feb 13 15:38:39.535601 kubelet[2499]: I0213 15:38:39.535447 2499 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4c8431e4-0aa5-49dd-b892-cbb23aae7a5d\" (UniqueName: \"kubernetes.io/nfs/a3cfabd3-e9ad-461e-b47b-eb3e70a258b1-pvc-4c8431e4-0aa5-49dd-b892-cbb23aae7a5d\") pod \"test-pod-1\" (UID: \"a3cfabd3-e9ad-461e-b47b-eb3e70a258b1\") " pod="default/test-pod-1" Feb 13 15:38:39.713179 kernel: FS-Cache: Loaded Feb 13 15:38:39.824726 kernel: RPC: Registered named UNIX socket transport module. Feb 13 15:38:39.825004 kernel: RPC: Registered udp transport module. Feb 13 15:38:39.825522 kernel: RPC: Registered tcp transport module. Feb 13 15:38:39.825596 kernel: RPC: Registered tcp-with-tls transport module. Feb 13 15:38:39.826140 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Feb 13 15:38:40.302166 kernel: NFS: Registering the id_resolver key type Feb 13 15:38:40.302285 kernel: Key type id_resolver registered Feb 13 15:38:40.302308 kernel: Key type id_legacy registered Feb 13 15:38:40.345314 nfsidmap[4652]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' Feb 13 15:38:40.350411 nfsidmap[4654]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' Feb 13 15:38:40.371225 kubelet[2499]: E0213 15:38:40.366830 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:40.625785 containerd[1992]: time="2025-02-13T15:38:40.625725231Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:a3cfabd3-e9ad-461e-b47b-eb3e70a258b1,Namespace:default,Attempt:0,}" Feb 13 15:38:40.918888 (udev-worker)[4653]: Network interface NamePolicy= disabled on kernel command line. Feb 13 15:38:40.923512 systemd-networkd[1568]: cali5ec59c6bf6e: Link UP Feb 13 15:38:40.925298 systemd-networkd[1568]: cali5ec59c6bf6e: Gained carrier Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.721 [INFO][4656] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.20.193-k8s-test--pod--1-eth0 default a3cfabd3-e9ad-461e-b47b-eb3e70a258b1 1314 0 2025-02-13 15:38:09 +0000 UTC map[projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.31.20.193 test-pod-1 eth0 default [] [] [kns.default ksa.default.default] cali5ec59c6bf6e [] []}} ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.722 [INFO][4656] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.776 [INFO][4666] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" HandleID="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Workload="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.793 [INFO][4666] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" HandleID="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Workload="172.31.20.193-k8s-test--pod--1-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003187b0), Attrs:map[string]string{"namespace":"default", "node":"172.31.20.193", "pod":"test-pod-1", "timestamp":"2025-02-13 15:38:40.776591549 +0000 UTC"}, Hostname:"172.31.20.193", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.796 [INFO][4666] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.798 [INFO][4666] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.798 [INFO][4666] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.20.193' Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.802 [INFO][4666] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.829 [INFO][4666] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.843 [INFO][4666] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.847 [INFO][4666] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.862 [INFO][4666] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.862 [INFO][4666] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.871 [INFO][4666] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063 Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.885 [INFO][4666] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.898 [INFO][4666] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.68/26] block=192.168.106.64/26 handle="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.898 [INFO][4666] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.68/26] handle="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" host="172.31.20.193" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.898 [INFO][4666] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.898 [INFO][4666] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.68/26] IPv6=[] ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" HandleID="k8s-pod-network.08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Workload="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.964683 containerd[1992]: 2025-02-13 15:38:40.902 [INFO][4656] cni-plugin/k8s.go 386: Populated endpoint ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"a3cfabd3-e9ad-461e-b47b-eb3e70a258b1", ResourceVersion:"1314", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.106.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:38:40.976242 containerd[1992]: 2025-02-13 15:38:40.902 [INFO][4656] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.68/32] ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.976242 containerd[1992]: 2025-02-13 15:38:40.902 [INFO][4656] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5ec59c6bf6e ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.976242 containerd[1992]: 2025-02-13 15:38:40.909 [INFO][4656] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:40.976242 containerd[1992]: 2025-02-13 15:38:40.912 [INFO][4656] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.20.193-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"a3cfabd3-e9ad-461e-b47b-eb3e70a258b1", ResourceVersion:"1314", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 15, 38, 9, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.20.193", ContainerID:"08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.106.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"62:46:dc:2f:a2:47", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 15:38:40.976242 containerd[1992]: 2025-02-13 15:38:40.960 [INFO][4656] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.20.193-k8s-test--pod--1-eth0" Feb 13 15:38:41.044939 containerd[1992]: time="2025-02-13T15:38:41.044727211Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 15:38:41.044939 containerd[1992]: time="2025-02-13T15:38:41.044867669Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 15:38:41.044939 containerd[1992]: time="2025-02-13T15:38:41.044895150Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:38:41.045298 containerd[1992]: time="2025-02-13T15:38:41.045011906Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 15:38:41.150333 containerd[1992]: time="2025-02-13T15:38:41.150210053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:a3cfabd3-e9ad-461e-b47b-eb3e70a258b1,Namespace:default,Attempt:0,} returns sandbox id \"08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063\"" Feb 13 15:38:41.168558 containerd[1992]: time="2025-02-13T15:38:41.168512947Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Feb 13 15:38:41.370788 kubelet[2499]: E0213 15:38:41.370655 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:41.589956 containerd[1992]: time="2025-02-13T15:38:41.589906906Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 15:38:41.592310 containerd[1992]: time="2025-02-13T15:38:41.592064704Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Feb 13 15:38:41.596295 containerd[1992]: time="2025-02-13T15:38:41.596247375Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:fe94eb5f0c9c8d0ca277aa8cd5940f1faf5970175bf373932babc578545deda8\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:d9bc3da999da9f147f1277c7b18292486847e8f39f95fcf81d914d0c22815faf\", size \"73054371\" in 427.690112ms" Feb 13 15:38:41.596295 containerd[1992]: time="2025-02-13T15:38:41.596288525Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fe94eb5f0c9c8d0ca277aa8cd5940f1faf5970175bf373932babc578545deda8\"" Feb 13 15:38:41.598322 containerd[1992]: time="2025-02-13T15:38:41.598284510Z" level=info msg="CreateContainer within sandbox \"08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063\" for container &ContainerMetadata{Name:test,Attempt:0,}" Feb 13 15:38:41.621570 containerd[1992]: time="2025-02-13T15:38:41.621386492Z" level=info msg="CreateContainer within sandbox \"08fa118f288a593ecd3ffc770edf72060586b116b7c6dfba8a61c9411b5d0063\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"e789b1705b8024774617c82cc5e81fa21b5abea81b985ff3b75df4131b0b58ee\"" Feb 13 15:38:41.622223 containerd[1992]: time="2025-02-13T15:38:41.622188067Z" level=info msg="StartContainer for \"e789b1705b8024774617c82cc5e81fa21b5abea81b985ff3b75df4131b0b58ee\"" Feb 13 15:38:41.705773 containerd[1992]: time="2025-02-13T15:38:41.705718148Z" level=info msg="StartContainer for \"e789b1705b8024774617c82cc5e81fa21b5abea81b985ff3b75df4131b0b58ee\" returns successfully" Feb 13 15:38:42.370972 kubelet[2499]: E0213 15:38:42.370915 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:42.458932 systemd-networkd[1568]: cali5ec59c6bf6e: Gained IPv6LL Feb 13 15:38:42.478665 kubelet[2499]: I0213 15:38:42.475289 2499 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=33.046779233 podStartE2EDuration="33.475244314s" podCreationTimestamp="2025-02-13 15:38:09 +0000 UTC" firstStartedPulling="2025-02-13 15:38:41.168135494 +0000 UTC m=+77.211780983" lastFinishedPulling="2025-02-13 15:38:41.596600579 +0000 UTC m=+77.640246064" observedRunningTime="2025-02-13 15:38:42.475111368 +0000 UTC m=+78.518756874" watchObservedRunningTime="2025-02-13 15:38:42.475244314 +0000 UTC m=+78.518889820" Feb 13 15:38:43.371236 kubelet[2499]: E0213 15:38:43.371188 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:44.282658 kubelet[2499]: E0213 15:38:44.282589 2499 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:44.372154 kubelet[2499]: E0213 15:38:44.372099 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:45.114160 ntpd[1955]: Listen normally on 11 cali5ec59c6bf6e [fe80::ecee:eeff:feee:eeee%9]:123 Feb 13 15:38:45.114621 ntpd[1955]: 13 Feb 15:38:45 ntpd[1955]: Listen normally on 11 cali5ec59c6bf6e [fe80::ecee:eeff:feee:eeee%9]:123 Feb 13 15:38:45.373243 kubelet[2499]: E0213 15:38:45.373101 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:46.373697 kubelet[2499]: E0213 15:38:46.373641 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:47.374847 kubelet[2499]: E0213 15:38:47.374787 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:48.375876 kubelet[2499]: E0213 15:38:48.375823 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:49.377062 kubelet[2499]: E0213 15:38:49.377002 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:50.377792 kubelet[2499]: E0213 15:38:50.377740 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:51.378933 kubelet[2499]: E0213 15:38:51.378879 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:52.379695 kubelet[2499]: E0213 15:38:52.379640 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:53.380611 kubelet[2499]: E0213 15:38:53.380561 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:54.381187 kubelet[2499]: E0213 15:38:54.381130 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:55.382006 kubelet[2499]: E0213 15:38:55.381952 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:56.382409 kubelet[2499]: E0213 15:38:56.382357 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:57.163771 kubelet[2499]: E0213 15:38:57.163720 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?resourceVersion=0&timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:38:57.174922 kubelet[2499]: E0213 15:38:57.174759 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:38:57.176156 kubelet[2499]: E0213 15:38:57.176036 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:38:57.176809 kubelet[2499]: E0213 15:38:57.176755 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:38:57.177408 kubelet[2499]: E0213 15:38:57.177383 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:38:57.177408 kubelet[2499]: E0213 15:38:57.177407 2499 kubelet_node_status.go:531] "Unable to update node status" err="update node status exceeds retry count" Feb 13 15:38:57.383238 kubelet[2499]: E0213 15:38:57.383129 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:58.384113 kubelet[2499]: E0213 15:38:58.384053 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:38:59.385180 kubelet[2499]: E0213 15:38:59.385027 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:00.385634 kubelet[2499]: E0213 15:39:00.385570 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:01.386229 kubelet[2499]: E0213 15:39:01.386170 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:02.386638 kubelet[2499]: E0213 15:39:02.386585 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:03.387393 kubelet[2499]: E0213 15:39:03.387337 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:04.282312 kubelet[2499]: E0213 15:39:04.282258 2499 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:04.388010 kubelet[2499]: E0213 15:39:04.387944 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:05.388935 kubelet[2499]: E0213 15:39:05.388877 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:06.389367 kubelet[2499]: E0213 15:39:06.389311 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:06.608669 kubelet[2499]: E0213 15:39:06.608631 2499 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:06.609294 kubelet[2499]: E0213 15:39:06.609261 2499 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:06.610209 kubelet[2499]: E0213 15:39:06.609725 2499 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:06.610477 kubelet[2499]: E0213 15:39:06.610447 2499 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:06.616404 kubelet[2499]: E0213 15:39:06.615654 2499 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:06.616404 kubelet[2499]: I0213 15:39:06.615696 2499 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 13 15:39:06.616626 kubelet[2499]: E0213 15:39:06.616521 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="200ms" Feb 13 15:39:06.817579 kubelet[2499]: E0213 15:39:06.817460 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="400ms" Feb 13 15:39:07.220115 kubelet[2499]: E0213 15:39:07.219558 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="800ms" Feb 13 15:39:07.254923 kubelet[2499]: E0213 15:39:07.254822 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?resourceVersion=0&timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:07.255963 kubelet[2499]: E0213 15:39:07.255797 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:07.256312 kubelet[2499]: E0213 15:39:07.256265 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:07.256991 kubelet[2499]: E0213 15:39:07.256846 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:07.257340 kubelet[2499]: E0213 15:39:07.257300 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:07.257468 kubelet[2499]: E0213 15:39:07.257317 2499 kubelet_node_status.go:531] "Unable to update node status" err="update node status exceeds retry count" Feb 13 15:39:07.390158 kubelet[2499]: E0213 15:39:07.390061 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:08.020687 kubelet[2499]: E0213 15:39:08.020645 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="1.6s" Feb 13 15:39:08.391257 kubelet[2499]: E0213 15:39:08.391212 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:09.392392 kubelet[2499]: E0213 15:39:09.392334 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:09.621575 kubelet[2499]: E0213 15:39:09.621535 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="3.2s" Feb 13 15:39:10.392817 kubelet[2499]: E0213 15:39:10.392762 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:11.393810 kubelet[2499]: E0213 15:39:11.393743 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:12.395006 kubelet[2499]: E0213 15:39:12.394943 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:12.822628 kubelet[2499]: E0213 15:39:12.822518 2499 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.86:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" interval="6.4s" Feb 13 15:39:13.395947 kubelet[2499]: E0213 15:39:13.395889 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:14.397062 kubelet[2499]: E0213 15:39:14.397002 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:15.398195 kubelet[2499]: E0213 15:39:15.398139 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:16.398499 kubelet[2499]: E0213 15:39:16.398441 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:17.398864 kubelet[2499]: E0213 15:39:17.398807 2499 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Feb 13 15:39:17.420032 kubelet[2499]: E0213 15:39:17.419988 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?resourceVersion=0&timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:17.420800 kubelet[2499]: E0213 15:39:17.420539 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:17.421042 kubelet[2499]: E0213 15:39:17.421019 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:17.422883 kubelet[2499]: E0213 15:39:17.422847 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:17.423664 kubelet[2499]: E0213 15:39:17.423639 2499 kubelet_node_status.go:544] "Error updating node status, will retry" err="error getting node \"172.31.20.193\": Get \"https://172.31.24.86:6443/api/v1/nodes/172.31.20.193?timeout=10s\": dial tcp 172.31.24.86:6443: connect: connection refused" Feb 13 15:39:17.423664 kubelet[2499]: E0213 15:39:17.423660 2499 kubelet_node_status.go:531] "Unable to update node status" err="update node status exceeds retry count"