Mar 17 17:25:45.223464 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Mar 17 17:25:45.223513 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT Mon Mar 17 16:05:23 -00 2025 Mar 17 17:25:45.223537 kernel: KASLR disabled due to lack of seed Mar 17 17:25:45.223555 kernel: efi: EFI v2.7 by EDK II Mar 17 17:25:45.223572 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7b003a98 MEMRESERVE=0x78503d98 Mar 17 17:25:45.223587 kernel: secureboot: Secure boot disabled Mar 17 17:25:45.223605 kernel: ACPI: Early table checksum verification disabled Mar 17 17:25:45.223620 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Mar 17 17:25:45.223636 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Mar 17 17:25:45.223651 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Mar 17 17:25:45.223672 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Mar 17 17:25:45.223688 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Mar 17 17:25:45.223703 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Mar 17 17:25:45.223719 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Mar 17 17:25:45.225514 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Mar 17 17:25:45.225557 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Mar 17 17:25:45.225576 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Mar 17 17:25:45.225593 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Mar 17 17:25:45.225611 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Mar 17 17:25:45.225628 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Mar 17 17:25:45.225645 kernel: printk: bootconsole [uart0] enabled Mar 17 17:25:45.225662 kernel: NUMA: Failed to initialise from firmware Mar 17 17:25:45.225680 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Mar 17 17:25:45.225696 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] Mar 17 17:25:45.225712 kernel: Zone ranges: Mar 17 17:25:45.225752 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Mar 17 17:25:45.225812 kernel: DMA32 empty Mar 17 17:25:45.225830 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Mar 17 17:25:45.225846 kernel: Movable zone start for each node Mar 17 17:25:45.225863 kernel: Early memory node ranges Mar 17 17:25:45.225880 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Mar 17 17:25:45.225896 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Mar 17 17:25:45.225913 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Mar 17 17:25:45.225930 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Mar 17 17:25:45.225946 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Mar 17 17:25:45.225962 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Mar 17 17:25:45.225978 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Mar 17 17:25:45.225994 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Mar 17 17:25:45.226018 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Mar 17 17:25:45.226050 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Mar 17 17:25:45.226115 kernel: psci: probing for conduit method from ACPI. Mar 17 17:25:45.226157 kernel: psci: PSCIv1.0 detected in firmware. Mar 17 17:25:45.226187 kernel: psci: Using standard PSCI v0.2 function IDs Mar 17 17:25:45.226214 kernel: psci: Trusted OS migration not required Mar 17 17:25:45.226235 kernel: psci: SMC Calling Convention v1.1 Mar 17 17:25:45.226254 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Mar 17 17:25:45.226272 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Mar 17 17:25:45.226290 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 17 17:25:45.226306 kernel: Detected PIPT I-cache on CPU0 Mar 17 17:25:45.226323 kernel: CPU features: detected: GIC system register CPU interface Mar 17 17:25:45.226340 kernel: CPU features: detected: Spectre-v2 Mar 17 17:25:45.226357 kernel: CPU features: detected: Spectre-v3a Mar 17 17:25:45.226374 kernel: CPU features: detected: Spectre-BHB Mar 17 17:25:45.226391 kernel: CPU features: detected: ARM erratum 1742098 Mar 17 17:25:45.226409 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Mar 17 17:25:45.226431 kernel: alternatives: applying boot alternatives Mar 17 17:25:45.226451 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=31b104f73129b84fa679201ebe02fbfd197d071bbf0576d6ccc5c5442bcbb405 Mar 17 17:25:45.226471 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 17 17:25:45.226488 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 17 17:25:45.226505 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 17 17:25:45.226522 kernel: Fallback order for Node 0: 0 Mar 17 17:25:45.226540 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Mar 17 17:25:45.226557 kernel: Policy zone: Normal Mar 17 17:25:45.226575 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 17 17:25:45.226592 kernel: software IO TLB: area num 2. Mar 17 17:25:45.226615 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Mar 17 17:25:45.226633 kernel: Memory: 3819896K/4030464K available (10240K kernel code, 2186K rwdata, 8100K rodata, 39744K init, 897K bss, 210568K reserved, 0K cma-reserved) Mar 17 17:25:45.226650 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 17 17:25:45.226668 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 17 17:25:45.226685 kernel: rcu: RCU event tracing is enabled. Mar 17 17:25:45.226703 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 17 17:25:45.226721 kernel: Trampoline variant of Tasks RCU enabled. Mar 17 17:25:45.226783 kernel: Tracing variant of Tasks RCU enabled. Mar 17 17:25:45.226801 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 17 17:25:45.226818 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 17 17:25:45.226835 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 17 17:25:45.226859 kernel: GICv3: 96 SPIs implemented Mar 17 17:25:45.226877 kernel: GICv3: 0 Extended SPIs implemented Mar 17 17:25:45.226894 kernel: Root IRQ handler: gic_handle_irq Mar 17 17:25:45.226911 kernel: GICv3: GICv3 features: 16 PPIs Mar 17 17:25:45.226928 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Mar 17 17:25:45.226945 kernel: ITS [mem 0x10080000-0x1009ffff] Mar 17 17:25:45.226962 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) Mar 17 17:25:45.226981 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) Mar 17 17:25:45.226998 kernel: GICv3: using LPI property table @0x00000004000d0000 Mar 17 17:25:45.227015 kernel: ITS: Using hypervisor restricted LPI range [128] Mar 17 17:25:45.227032 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 Mar 17 17:25:45.227049 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 17 17:25:45.227070 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Mar 17 17:25:45.227087 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Mar 17 17:25:45.227105 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Mar 17 17:25:45.227122 kernel: Console: colour dummy device 80x25 Mar 17 17:25:45.227139 kernel: printk: console [tty1] enabled Mar 17 17:25:45.227157 kernel: ACPI: Core revision 20230628 Mar 17 17:25:45.227175 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Mar 17 17:25:45.227192 kernel: pid_max: default: 32768 minimum: 301 Mar 17 17:25:45.227210 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 17 17:25:45.227227 kernel: landlock: Up and running. Mar 17 17:25:45.227249 kernel: SELinux: Initializing. Mar 17 17:25:45.227267 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 17:25:45.227284 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 17:25:45.227302 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 17 17:25:45.227320 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 17 17:25:45.227337 kernel: rcu: Hierarchical SRCU implementation. Mar 17 17:25:45.227355 kernel: rcu: Max phase no-delay instances is 400. Mar 17 17:25:45.227373 kernel: Platform MSI: ITS@0x10080000 domain created Mar 17 17:25:45.227394 kernel: PCI/MSI: ITS@0x10080000 domain created Mar 17 17:25:45.227412 kernel: Remapping and enabling EFI services. Mar 17 17:25:45.227429 kernel: smp: Bringing up secondary CPUs ... Mar 17 17:25:45.227447 kernel: Detected PIPT I-cache on CPU1 Mar 17 17:25:45.227464 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Mar 17 17:25:45.227482 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 Mar 17 17:25:45.227499 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Mar 17 17:25:45.227516 kernel: smp: Brought up 1 node, 2 CPUs Mar 17 17:25:45.227534 kernel: SMP: Total of 2 processors activated. Mar 17 17:25:45.227551 kernel: CPU features: detected: 32-bit EL0 Support Mar 17 17:25:45.227573 kernel: CPU features: detected: 32-bit EL1 Support Mar 17 17:25:45.227591 kernel: CPU features: detected: CRC32 instructions Mar 17 17:25:45.227619 kernel: CPU: All CPU(s) started at EL1 Mar 17 17:25:45.227642 kernel: alternatives: applying system-wide alternatives Mar 17 17:25:45.227660 kernel: devtmpfs: initialized Mar 17 17:25:45.227678 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 17 17:25:45.227697 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 17 17:25:45.227716 kernel: pinctrl core: initialized pinctrl subsystem Mar 17 17:25:45.227762 kernel: SMBIOS 3.0.0 present. Mar 17 17:25:45.227790 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Mar 17 17:25:45.227808 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 17 17:25:45.227826 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 17 17:25:45.227845 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 17 17:25:45.227863 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 17 17:25:45.227882 kernel: audit: initializing netlink subsys (disabled) Mar 17 17:25:45.227900 kernel: audit: type=2000 audit(0.232:1): state=initialized audit_enabled=0 res=1 Mar 17 17:25:45.227922 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 17 17:25:45.227941 kernel: cpuidle: using governor menu Mar 17 17:25:45.227959 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 17 17:25:45.227977 kernel: ASID allocator initialised with 65536 entries Mar 17 17:25:45.227995 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 17 17:25:45.228014 kernel: Serial: AMBA PL011 UART driver Mar 17 17:25:45.228033 kernel: Modules: 17424 pages in range for non-PLT usage Mar 17 17:25:45.228051 kernel: Modules: 508944 pages in range for PLT usage Mar 17 17:25:45.228069 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 17 17:25:45.228092 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 17 17:25:45.228110 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 17 17:25:45.228128 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 17 17:25:45.228146 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 17 17:25:45.228165 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 17 17:25:45.228182 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 17 17:25:45.228200 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 17 17:25:45.228218 kernel: ACPI: Added _OSI(Module Device) Mar 17 17:25:45.228236 kernel: ACPI: Added _OSI(Processor Device) Mar 17 17:25:45.228258 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 17 17:25:45.228277 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 17 17:25:45.228295 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 17 17:25:45.228313 kernel: ACPI: Interpreter enabled Mar 17 17:25:45.228331 kernel: ACPI: Using GIC for interrupt routing Mar 17 17:25:45.228349 kernel: ACPI: MCFG table detected, 1 entries Mar 17 17:25:45.228367 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Mar 17 17:25:45.228719 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 17 17:25:45.229027 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Mar 17 17:25:45.229237 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Mar 17 17:25:45.229480 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Mar 17 17:25:45.229705 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Mar 17 17:25:45.229781 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Mar 17 17:25:45.229803 kernel: acpiphp: Slot [1] registered Mar 17 17:25:45.229822 kernel: acpiphp: Slot [2] registered Mar 17 17:25:45.229840 kernel: acpiphp: Slot [3] registered Mar 17 17:25:45.229868 kernel: acpiphp: Slot [4] registered Mar 17 17:25:45.229887 kernel: acpiphp: Slot [5] registered Mar 17 17:25:45.229905 kernel: acpiphp: Slot [6] registered Mar 17 17:25:45.229924 kernel: acpiphp: Slot [7] registered Mar 17 17:25:45.229942 kernel: acpiphp: Slot [8] registered Mar 17 17:25:45.229960 kernel: acpiphp: Slot [9] registered Mar 17 17:25:45.229978 kernel: acpiphp: Slot [10] registered Mar 17 17:25:45.229996 kernel: acpiphp: Slot [11] registered Mar 17 17:25:45.230014 kernel: acpiphp: Slot [12] registered Mar 17 17:25:45.230032 kernel: acpiphp: Slot [13] registered Mar 17 17:25:45.230055 kernel: acpiphp: Slot [14] registered Mar 17 17:25:45.230073 kernel: acpiphp: Slot [15] registered Mar 17 17:25:45.230091 kernel: acpiphp: Slot [16] registered Mar 17 17:25:45.230109 kernel: acpiphp: Slot [17] registered Mar 17 17:25:45.230127 kernel: acpiphp: Slot [18] registered Mar 17 17:25:45.230145 kernel: acpiphp: Slot [19] registered Mar 17 17:25:45.230164 kernel: acpiphp: Slot [20] registered Mar 17 17:25:45.230182 kernel: acpiphp: Slot [21] registered Mar 17 17:25:45.230200 kernel: acpiphp: Slot [22] registered Mar 17 17:25:45.230223 kernel: acpiphp: Slot [23] registered Mar 17 17:25:45.230241 kernel: acpiphp: Slot [24] registered Mar 17 17:25:45.230258 kernel: acpiphp: Slot [25] registered Mar 17 17:25:45.230276 kernel: acpiphp: Slot [26] registered Mar 17 17:25:45.230294 kernel: acpiphp: Slot [27] registered Mar 17 17:25:45.230312 kernel: acpiphp: Slot [28] registered Mar 17 17:25:45.230330 kernel: acpiphp: Slot [29] registered Mar 17 17:25:45.230348 kernel: acpiphp: Slot [30] registered Mar 17 17:25:45.230366 kernel: acpiphp: Slot [31] registered Mar 17 17:25:45.230383 kernel: PCI host bridge to bus 0000:00 Mar 17 17:25:45.230656 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Mar 17 17:25:45.230957 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Mar 17 17:25:45.231161 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Mar 17 17:25:45.233707 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Mar 17 17:25:45.234092 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Mar 17 17:25:45.234387 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Mar 17 17:25:45.234677 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Mar 17 17:25:45.235027 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Mar 17 17:25:45.235271 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Mar 17 17:25:45.238103 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Mar 17 17:25:45.238382 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Mar 17 17:25:45.238614 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Mar 17 17:25:45.238907 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Mar 17 17:25:45.239154 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Mar 17 17:25:45.239378 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Mar 17 17:25:45.239883 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Mar 17 17:25:45.240140 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Mar 17 17:25:45.240365 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Mar 17 17:25:45.240583 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Mar 17 17:25:45.240888 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Mar 17 17:25:45.241119 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Mar 17 17:25:45.241339 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Mar 17 17:25:45.241547 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Mar 17 17:25:45.241578 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Mar 17 17:25:45.241598 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Mar 17 17:25:45.241617 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Mar 17 17:25:45.241636 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Mar 17 17:25:45.241654 kernel: iommu: Default domain type: Translated Mar 17 17:25:45.241682 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 17 17:25:45.241701 kernel: efivars: Registered efivars operations Mar 17 17:25:45.241719 kernel: vgaarb: loaded Mar 17 17:25:45.241782 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 17 17:25:45.241803 kernel: VFS: Disk quotas dquot_6.6.0 Mar 17 17:25:45.241823 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 17 17:25:45.241842 kernel: pnp: PnP ACPI init Mar 17 17:25:45.242111 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Mar 17 17:25:45.242154 kernel: pnp: PnP ACPI: found 1 devices Mar 17 17:25:45.242174 kernel: NET: Registered PF_INET protocol family Mar 17 17:25:45.242193 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 17 17:25:45.242211 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 17 17:25:45.242230 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 17 17:25:45.242248 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 17 17:25:45.242267 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 17 17:25:45.242288 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 17 17:25:45.242306 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 17:25:45.242331 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 17:25:45.242350 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 17 17:25:45.242368 kernel: PCI: CLS 0 bytes, default 64 Mar 17 17:25:45.242387 kernel: kvm [1]: HYP mode not available Mar 17 17:25:45.242405 kernel: Initialise system trusted keyrings Mar 17 17:25:45.242424 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 17 17:25:45.242444 kernel: Key type asymmetric registered Mar 17 17:25:45.242462 kernel: Asymmetric key parser 'x509' registered Mar 17 17:25:45.242480 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 17 17:25:45.242502 kernel: io scheduler mq-deadline registered Mar 17 17:25:45.242521 kernel: io scheduler kyber registered Mar 17 17:25:45.242539 kernel: io scheduler bfq registered Mar 17 17:25:45.242842 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Mar 17 17:25:45.242879 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Mar 17 17:25:45.242899 kernel: ACPI: button: Power Button [PWRB] Mar 17 17:25:45.242918 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Mar 17 17:25:45.242936 kernel: ACPI: button: Sleep Button [SLPB] Mar 17 17:25:45.242966 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 17 17:25:45.242987 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Mar 17 17:25:45.243238 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Mar 17 17:25:45.243272 kernel: printk: console [ttyS0] disabled Mar 17 17:25:45.243292 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Mar 17 17:25:45.243312 kernel: printk: console [ttyS0] enabled Mar 17 17:25:45.243331 kernel: printk: bootconsole [uart0] disabled Mar 17 17:25:45.243351 kernel: thunder_xcv, ver 1.0 Mar 17 17:25:45.243370 kernel: thunder_bgx, ver 1.0 Mar 17 17:25:45.243390 kernel: nicpf, ver 1.0 Mar 17 17:25:45.243418 kernel: nicvf, ver 1.0 Mar 17 17:25:45.243701 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 17 17:25:45.243986 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-03-17T17:25:44 UTC (1742232344) Mar 17 17:25:45.244015 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 17 17:25:45.244034 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Mar 17 17:25:45.244053 kernel: watchdog: Delayed init of the lockup detector failed: -19 Mar 17 17:25:45.244072 kernel: watchdog: Hard watchdog permanently disabled Mar 17 17:25:45.244100 kernel: NET: Registered PF_INET6 protocol family Mar 17 17:25:45.244119 kernel: Segment Routing with IPv6 Mar 17 17:25:45.244138 kernel: In-situ OAM (IOAM) with IPv6 Mar 17 17:25:45.244156 kernel: NET: Registered PF_PACKET protocol family Mar 17 17:25:45.244174 kernel: Key type dns_resolver registered Mar 17 17:25:45.244193 kernel: registered taskstats version 1 Mar 17 17:25:45.244211 kernel: Loading compiled-in X.509 certificates Mar 17 17:25:45.244230 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: 74c9b4f5dfad711856d7363c976664fc02c1e24c' Mar 17 17:25:45.244249 kernel: Key type .fscrypt registered Mar 17 17:25:45.244267 kernel: Key type fscrypt-provisioning registered Mar 17 17:25:45.244292 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 17 17:25:45.244310 kernel: ima: Allocated hash algorithm: sha1 Mar 17 17:25:45.244329 kernel: ima: No architecture policies found Mar 17 17:25:45.244347 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 17 17:25:45.244365 kernel: clk: Disabling unused clocks Mar 17 17:25:45.244385 kernel: Freeing unused kernel memory: 39744K Mar 17 17:25:45.244403 kernel: Run /init as init process Mar 17 17:25:45.244422 kernel: with arguments: Mar 17 17:25:45.244441 kernel: /init Mar 17 17:25:45.244465 kernel: with environment: Mar 17 17:25:45.244484 kernel: HOME=/ Mar 17 17:25:45.244504 kernel: TERM=linux Mar 17 17:25:45.244521 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 17 17:25:45.244544 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 17 17:25:45.244568 systemd[1]: Detected virtualization amazon. Mar 17 17:25:45.244589 systemd[1]: Detected architecture arm64. Mar 17 17:25:45.244613 systemd[1]: Running in initrd. Mar 17 17:25:45.244634 systemd[1]: No hostname configured, using default hostname. Mar 17 17:25:45.244653 systemd[1]: Hostname set to . Mar 17 17:25:45.244673 systemd[1]: Initializing machine ID from VM UUID. Mar 17 17:25:45.244692 systemd[1]: Queued start job for default target initrd.target. Mar 17 17:25:45.244713 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:25:45.244766 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:25:45.244791 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 17 17:25:45.244819 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 17 17:25:45.244840 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 17 17:25:45.244861 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 17 17:25:45.244884 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 17 17:25:45.244904 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 17 17:25:45.244924 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:25:45.244944 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:25:45.244970 systemd[1]: Reached target paths.target - Path Units. Mar 17 17:25:45.244993 systemd[1]: Reached target slices.target - Slice Units. Mar 17 17:25:45.245012 systemd[1]: Reached target swap.target - Swaps. Mar 17 17:25:45.245032 systemd[1]: Reached target timers.target - Timer Units. Mar 17 17:25:45.245052 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 17 17:25:45.245072 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 17 17:25:45.245092 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 17 17:25:45.245112 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 17 17:25:45.245132 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:25:45.245157 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 17 17:25:45.245177 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:25:45.245197 systemd[1]: Reached target sockets.target - Socket Units. Mar 17 17:25:45.245217 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 17 17:25:45.245237 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 17 17:25:45.245257 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 17 17:25:45.245277 systemd[1]: Starting systemd-fsck-usr.service... Mar 17 17:25:45.245297 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 17 17:25:45.245351 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 17 17:25:45.245372 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:25:45.245392 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 17 17:25:45.245463 systemd-journald[251]: Collecting audit messages is disabled. Mar 17 17:25:45.245514 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:25:45.245536 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 17 17:25:45.245557 systemd[1]: Finished systemd-fsck-usr.service. Mar 17 17:25:45.245578 kernel: Bridge firewalling registered Mar 17 17:25:45.245598 systemd-journald[251]: Journal started Mar 17 17:25:45.245641 systemd-journald[251]: Runtime Journal (/run/log/journal/ec2786e69a82c3ce72a6a6a1720d632d) is 8.0M, max 75.3M, 67.3M free. Mar 17 17:25:45.201974 systemd-modules-load[252]: Inserted module 'overlay' Mar 17 17:25:45.235430 systemd-modules-load[252]: Inserted module 'br_netfilter' Mar 17 17:25:45.260774 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 17 17:25:45.260864 systemd[1]: Started systemd-journald.service - Journal Service. Mar 17 17:25:45.265122 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 17 17:25:45.274214 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:25:45.288147 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:25:45.301056 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 17 17:25:45.307085 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 17 17:25:45.307708 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 17 17:25:45.316172 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 17 17:25:45.353008 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:25:45.366028 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:25:45.375861 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:25:45.388995 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 17 17:25:45.396219 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:25:45.409034 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 17 17:25:45.467035 dracut-cmdline[289]: dracut-dracut-053 Mar 17 17:25:45.471997 systemd-resolved[287]: Positive Trust Anchors: Mar 17 17:25:45.473804 systemd-resolved[287]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 17:25:45.478909 dracut-cmdline[289]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=31b104f73129b84fa679201ebe02fbfd197d071bbf0576d6ccc5c5442bcbb405 Mar 17 17:25:45.473870 systemd-resolved[287]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 17 17:25:45.614768 kernel: SCSI subsystem initialized Mar 17 17:25:45.624761 kernel: Loading iSCSI transport class v2.0-870. Mar 17 17:25:45.635775 kernel: iscsi: registered transport (tcp) Mar 17 17:25:45.657992 kernel: iscsi: registered transport (qla4xxx) Mar 17 17:25:45.658068 kernel: QLogic iSCSI HBA Driver Mar 17 17:25:45.727781 kernel: random: crng init done Mar 17 17:25:45.728063 systemd-resolved[287]: Defaulting to hostname 'linux'. Mar 17 17:25:45.731692 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 17 17:25:45.735703 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:25:45.762823 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 17 17:25:45.778680 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 17 17:25:45.812168 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 17 17:25:45.812247 kernel: device-mapper: uevent: version 1.0.3 Mar 17 17:25:45.813952 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 17 17:25:45.882811 kernel: raid6: neonx8 gen() 6722 MB/s Mar 17 17:25:45.899787 kernel: raid6: neonx4 gen() 6479 MB/s Mar 17 17:25:45.916772 kernel: raid6: neonx2 gen() 5430 MB/s Mar 17 17:25:45.933780 kernel: raid6: neonx1 gen() 3950 MB/s Mar 17 17:25:45.950774 kernel: raid6: int64x8 gen() 3826 MB/s Mar 17 17:25:45.967778 kernel: raid6: int64x4 gen() 3726 MB/s Mar 17 17:25:45.984769 kernel: raid6: int64x2 gen() 3607 MB/s Mar 17 17:25:46.002665 kernel: raid6: int64x1 gen() 2737 MB/s Mar 17 17:25:46.002769 kernel: raid6: using algorithm neonx8 gen() 6722 MB/s Mar 17 17:25:46.020633 kernel: raid6: .... xor() 4765 MB/s, rmw enabled Mar 17 17:25:46.020707 kernel: raid6: using neon recovery algorithm Mar 17 17:25:46.030063 kernel: xor: measuring software checksum speed Mar 17 17:25:46.030143 kernel: 8regs : 11019 MB/sec Mar 17 17:25:46.031224 kernel: 32regs : 11894 MB/sec Mar 17 17:25:46.032437 kernel: arm64_neon : 9440 MB/sec Mar 17 17:25:46.032500 kernel: xor: using function: 32regs (11894 MB/sec) Mar 17 17:25:46.120786 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 17 17:25:46.143276 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 17 17:25:46.157060 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:25:46.198415 systemd-udevd[470]: Using default interface naming scheme 'v255'. Mar 17 17:25:46.208679 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:25:46.221057 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 17 17:25:46.263508 dracut-pre-trigger[475]: rd.md=0: removing MD RAID activation Mar 17 17:25:46.325707 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 17 17:25:46.336069 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 17 17:25:46.470411 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:25:46.487256 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 17 17:25:46.533576 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 17 17:25:46.540242 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 17 17:25:46.545878 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:25:46.548419 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 17 17:25:46.570072 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 17 17:25:46.611241 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 17 17:25:46.695828 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Mar 17 17:25:46.695917 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Mar 17 17:25:46.730278 kernel: ena 0000:00:05.0: ENA device version: 0.10 Mar 17 17:25:46.730584 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Mar 17 17:25:46.730943 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:6c:a0:0a:34:99 Mar 17 17:25:46.731194 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Mar 17 17:25:46.702863 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 17:25:46.703137 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:25:46.741389 kernel: nvme nvme0: pci function 0000:00:04.0 Mar 17 17:25:46.709156 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:25:46.713594 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:25:46.713928 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:25:46.719960 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:25:46.741552 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:25:46.775786 kernel: nvme nvme0: 2/0/0 default/read/poll queues Mar 17 17:25:46.779230 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:25:46.789130 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:25:46.799310 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 17 17:25:46.799387 kernel: GPT:9289727 != 16777215 Mar 17 17:25:46.799414 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 17 17:25:46.799439 kernel: GPT:9289727 != 16777215 Mar 17 17:25:46.800509 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 17 17:25:46.800580 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Mar 17 17:25:46.805645 (udev-worker)[514]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:25:46.841970 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:25:46.900776 kernel: BTRFS: device fsid c0c482e3-6885-4a4e-b31c-6bc8f8c403e7 devid 1 transid 40 /dev/nvme0n1p3 scanned by (udev-worker) (547) Mar 17 17:25:46.932799 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by (udev-worker) (515) Mar 17 17:25:46.956823 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Mar 17 17:25:47.017382 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Mar 17 17:25:47.075522 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Mar 17 17:25:47.091351 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Mar 17 17:25:47.094872 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Mar 17 17:25:47.113034 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 17 17:25:47.125439 disk-uuid[660]: Primary Header is updated. Mar 17 17:25:47.125439 disk-uuid[660]: Secondary Entries is updated. Mar 17 17:25:47.125439 disk-uuid[660]: Secondary Header is updated. Mar 17 17:25:47.136768 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Mar 17 17:25:48.153825 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Mar 17 17:25:48.155147 disk-uuid[661]: The operation has completed successfully. Mar 17 17:25:48.359706 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 17 17:25:48.359994 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 17 17:25:48.413087 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 17 17:25:48.424135 sh[922]: Success Mar 17 17:25:48.452022 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 17 17:25:48.565443 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 17 17:25:48.570604 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 17 17:25:48.584917 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 17 17:25:48.623818 kernel: BTRFS info (device dm-0): first mount of filesystem c0c482e3-6885-4a4e-b31c-6bc8f8c403e7 Mar 17 17:25:48.623885 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 17 17:25:48.623923 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 17 17:25:48.626972 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 17 17:25:48.627018 kernel: BTRFS info (device dm-0): using free space tree Mar 17 17:25:48.656772 kernel: BTRFS info (device dm-0): enabling ssd optimizations Mar 17 17:25:48.673899 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 17 17:25:48.678028 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 17 17:25:48.690017 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 17 17:25:48.697034 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 17 17:25:48.732014 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 3dbd9b64-bd31-4292-be10-51551993b53f Mar 17 17:25:48.732090 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Mar 17 17:25:48.732129 kernel: BTRFS info (device nvme0n1p6): using free space tree Mar 17 17:25:48.741795 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Mar 17 17:25:48.760358 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 17 17:25:48.763472 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 3dbd9b64-bd31-4292-be10-51551993b53f Mar 17 17:25:48.773796 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 17 17:25:48.786146 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 17 17:25:48.885911 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 17 17:25:48.902176 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 17 17:25:48.960145 systemd-networkd[1115]: lo: Link UP Mar 17 17:25:48.960166 systemd-networkd[1115]: lo: Gained carrier Mar 17 17:25:48.965364 systemd-networkd[1115]: Enumeration completed Mar 17 17:25:48.966176 systemd-networkd[1115]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:25:48.966184 systemd-networkd[1115]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 17:25:48.966710 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 17 17:25:48.980207 systemd[1]: Reached target network.target - Network. Mar 17 17:25:48.985907 systemd-networkd[1115]: eth0: Link UP Mar 17 17:25:48.985916 systemd-networkd[1115]: eth0: Gained carrier Mar 17 17:25:48.985934 systemd-networkd[1115]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:25:49.007871 systemd-networkd[1115]: eth0: DHCPv4 address 172.31.31.32/20, gateway 172.31.16.1 acquired from 172.31.16.1 Mar 17 17:25:49.117671 ignition[1034]: Ignition 2.20.0 Mar 17 17:25:49.117695 ignition[1034]: Stage: fetch-offline Mar 17 17:25:49.118199 ignition[1034]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:49.118227 ignition[1034]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:49.120812 ignition[1034]: Ignition finished successfully Mar 17 17:25:49.125573 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 17 17:25:49.150186 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 17 17:25:49.174568 ignition[1124]: Ignition 2.20.0 Mar 17 17:25:49.174606 ignition[1124]: Stage: fetch Mar 17 17:25:49.175327 ignition[1124]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:49.175357 ignition[1124]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:49.175559 ignition[1124]: PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:49.187326 ignition[1124]: PUT result: OK Mar 17 17:25:49.190613 ignition[1124]: parsed url from cmdline: "" Mar 17 17:25:49.190635 ignition[1124]: no config URL provided Mar 17 17:25:49.190654 ignition[1124]: reading system config file "/usr/lib/ignition/user.ign" Mar 17 17:25:49.190682 ignition[1124]: no config at "/usr/lib/ignition/user.ign" Mar 17 17:25:49.190720 ignition[1124]: PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:49.192676 ignition[1124]: PUT result: OK Mar 17 17:25:49.205114 unknown[1124]: fetched base config from "system" Mar 17 17:25:49.192784 ignition[1124]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Mar 17 17:25:49.205131 unknown[1124]: fetched base config from "system" Mar 17 17:25:49.195222 ignition[1124]: GET result: OK Mar 17 17:25:49.205144 unknown[1124]: fetched user config from "aws" Mar 17 17:25:49.195336 ignition[1124]: parsing config with SHA512: 2a41ccf4237706f5c8e9e04d25947c44f108c9c7c421265766b2d0966fc4e71428819992d327217be97d6e923b462de0002761073d3fe14da4656f74c3dc99b5 Mar 17 17:25:49.214815 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 17 17:25:49.205689 ignition[1124]: fetch: fetch complete Mar 17 17:25:49.205702 ignition[1124]: fetch: fetch passed Mar 17 17:25:49.205818 ignition[1124]: Ignition finished successfully Mar 17 17:25:49.233425 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 17 17:25:49.260512 ignition[1131]: Ignition 2.20.0 Mar 17 17:25:49.260557 ignition[1131]: Stage: kargs Mar 17 17:25:49.261975 ignition[1131]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:49.262002 ignition[1131]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:49.262234 ignition[1131]: PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:49.268058 ignition[1131]: PUT result: OK Mar 17 17:25:49.276601 ignition[1131]: kargs: kargs passed Mar 17 17:25:49.276812 ignition[1131]: Ignition finished successfully Mar 17 17:25:49.282020 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 17 17:25:49.291056 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 17 17:25:49.324914 ignition[1138]: Ignition 2.20.0 Mar 17 17:25:49.324951 ignition[1138]: Stage: disks Mar 17 17:25:49.326655 ignition[1138]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:49.326684 ignition[1138]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:49.327853 ignition[1138]: PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:49.334137 ignition[1138]: PUT result: OK Mar 17 17:25:49.339017 ignition[1138]: disks: disks passed Mar 17 17:25:49.339165 ignition[1138]: Ignition finished successfully Mar 17 17:25:49.345806 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 17 17:25:49.349009 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 17 17:25:49.351428 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 17 17:25:49.351711 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 17 17:25:49.352307 systemd[1]: Reached target sysinit.target - System Initialization. Mar 17 17:25:49.352606 systemd[1]: Reached target basic.target - Basic System. Mar 17 17:25:49.383132 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 17 17:25:49.434846 systemd-fsck[1146]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 17 17:25:49.444292 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 17 17:25:49.452980 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 17 17:25:49.551788 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 6b579bf2-7716-4d59-98eb-b92ea668693e r/w with ordered data mode. Quota mode: none. Mar 17 17:25:49.553583 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 17 17:25:49.557469 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 17 17:25:49.579942 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 17 17:25:49.585999 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 17 17:25:49.588474 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 17 17:25:49.588557 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 17 17:25:49.588613 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 17 17:25:49.609506 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1165) Mar 17 17:25:49.614091 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 3dbd9b64-bd31-4292-be10-51551993b53f Mar 17 17:25:49.614166 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Mar 17 17:25:49.615338 kernel: BTRFS info (device nvme0n1p6): using free space tree Mar 17 17:25:49.623505 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 17 17:25:49.632794 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Mar 17 17:25:49.633316 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 17 17:25:49.650310 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 17 17:25:49.802851 initrd-setup-root[1190]: cut: /sysroot/etc/passwd: No such file or directory Mar 17 17:25:49.813164 initrd-setup-root[1197]: cut: /sysroot/etc/group: No such file or directory Mar 17 17:25:49.821562 initrd-setup-root[1204]: cut: /sysroot/etc/shadow: No such file or directory Mar 17 17:25:49.831915 initrd-setup-root[1211]: cut: /sysroot/etc/gshadow: No such file or directory Mar 17 17:25:49.992615 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 17 17:25:50.001946 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 17 17:25:50.007855 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 17 17:25:50.031533 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 17 17:25:50.033816 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 3dbd9b64-bd31-4292-be10-51551993b53f Mar 17 17:25:50.067768 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 17 17:25:50.080917 ignition[1279]: INFO : Ignition 2.20.0 Mar 17 17:25:50.080917 ignition[1279]: INFO : Stage: mount Mar 17 17:25:50.084052 ignition[1279]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:50.084052 ignition[1279]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:50.084052 ignition[1279]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:50.090945 ignition[1279]: INFO : PUT result: OK Mar 17 17:25:50.094638 ignition[1279]: INFO : mount: mount passed Mar 17 17:25:50.094638 ignition[1279]: INFO : Ignition finished successfully Mar 17 17:25:50.099121 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 17 17:25:50.114080 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 17 17:25:50.129371 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 17 17:25:50.165753 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by mount (1290) Mar 17 17:25:50.170492 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 3dbd9b64-bd31-4292-be10-51551993b53f Mar 17 17:25:50.170570 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Mar 17 17:25:50.170601 kernel: BTRFS info (device nvme0n1p6): using free space tree Mar 17 17:25:50.175789 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Mar 17 17:25:50.179303 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 17 17:25:50.216087 ignition[1307]: INFO : Ignition 2.20.0 Mar 17 17:25:50.216087 ignition[1307]: INFO : Stage: files Mar 17 17:25:50.219462 ignition[1307]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:50.219462 ignition[1307]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:50.219462 ignition[1307]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:50.226599 ignition[1307]: INFO : PUT result: OK Mar 17 17:25:50.230541 ignition[1307]: DEBUG : files: compiled without relabeling support, skipping Mar 17 17:25:50.234072 ignition[1307]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 17 17:25:50.234072 ignition[1307]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 17 17:25:50.244337 ignition[1307]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 17 17:25:50.246980 ignition[1307]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 17 17:25:50.250126 unknown[1307]: wrote ssh authorized keys file for user: core Mar 17 17:25:50.253879 ignition[1307]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 17 17:25:50.258820 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 17 17:25:50.258820 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Mar 17 17:25:50.258820 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 17 17:25:50.258820 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 17:25:50.273431 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Mar 17 17:25:50.742472 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Mar 17 17:25:50.832910 systemd-networkd[1115]: eth0: Gained IPv6LL Mar 17 17:25:51.144844 ignition[1307]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 17:25:51.144844 ignition[1307]: INFO : files: op(8): [started] processing unit "containerd.service" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: op(8): op(9): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: op(8): op(9): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: op(8): [finished] processing unit "containerd.service" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: createResultFile: createFiles: op(a): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: createResultFile: createFiles: op(a): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 17 17:25:51.151980 ignition[1307]: INFO : files: files passed Mar 17 17:25:51.151980 ignition[1307]: INFO : Ignition finished successfully Mar 17 17:25:51.175116 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 17 17:25:51.192034 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 17 17:25:51.198461 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 17 17:25:51.208479 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 17 17:25:51.208834 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 17 17:25:51.245181 initrd-setup-root-after-ignition[1336]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:25:51.249123 initrd-setup-root-after-ignition[1340]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:25:51.252264 initrd-setup-root-after-ignition[1336]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:25:51.258810 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 17 17:25:51.259420 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 17 17:25:51.286198 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 17 17:25:51.339868 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 17 17:25:51.340918 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 17 17:25:51.347769 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 17 17:25:51.349859 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 17 17:25:51.351957 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 17 17:25:51.363073 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 17 17:25:51.394460 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 17 17:25:51.415167 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 17 17:25:51.440900 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:25:51.443868 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:25:51.450880 systemd[1]: Stopped target timers.target - Timer Units. Mar 17 17:25:51.452955 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 17 17:25:51.453213 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 17 17:25:51.456267 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 17 17:25:51.466343 systemd[1]: Stopped target basic.target - Basic System. Mar 17 17:25:51.468722 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 17 17:25:51.473968 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 17 17:25:51.478379 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 17 17:25:51.480768 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 17 17:25:51.483129 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 17 17:25:51.491391 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 17 17:25:51.493623 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 17 17:25:51.500525 systemd[1]: Stopped target swap.target - Swaps. Mar 17 17:25:51.503047 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 17 17:25:51.503288 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 17 17:25:51.510128 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:25:51.512856 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:25:51.518714 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 17 17:25:51.521406 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:25:51.524135 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 17 17:25:51.524374 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 17 17:25:51.532951 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 17 17:25:51.533225 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 17 17:25:51.535784 systemd[1]: ignition-files.service: Deactivated successfully. Mar 17 17:25:51.535988 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 17 17:25:51.552188 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 17 17:25:51.555107 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 17 17:25:51.555375 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:25:51.578868 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 17 17:25:51.582596 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 17 17:25:51.584961 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:25:51.589593 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 17 17:25:51.589852 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 17 17:25:51.608650 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 17 17:25:51.608905 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 17 17:25:51.626138 ignition[1360]: INFO : Ignition 2.20.0 Mar 17 17:25:51.628567 ignition[1360]: INFO : Stage: umount Mar 17 17:25:51.630377 ignition[1360]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:25:51.630377 ignition[1360]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Mar 17 17:25:51.630377 ignition[1360]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Mar 17 17:25:51.639339 ignition[1360]: INFO : PUT result: OK Mar 17 17:25:51.643277 ignition[1360]: INFO : umount: umount passed Mar 17 17:25:51.645109 ignition[1360]: INFO : Ignition finished successfully Mar 17 17:25:51.649519 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 17 17:25:51.649831 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 17 17:25:51.653609 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 17 17:25:51.657528 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 17 17:25:51.661531 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 17 17:25:51.661650 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 17 17:25:51.663632 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 17 17:25:51.663723 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 17 17:25:51.665709 systemd[1]: Stopped target network.target - Network. Mar 17 17:25:51.667422 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 17 17:25:51.667544 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 17 17:25:51.669932 systemd[1]: Stopped target paths.target - Path Units. Mar 17 17:25:51.671647 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 17 17:25:51.676083 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:25:51.694671 systemd[1]: Stopped target slices.target - Slice Units. Mar 17 17:25:51.696545 systemd[1]: Stopped target sockets.target - Socket Units. Mar 17 17:25:51.699222 systemd[1]: iscsid.socket: Deactivated successfully. Mar 17 17:25:51.699310 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 17 17:25:51.704836 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 17 17:25:51.704928 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 17 17:25:51.708174 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 17 17:25:51.708282 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 17 17:25:51.722315 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 17 17:25:51.722495 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 17 17:25:51.725156 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 17 17:25:51.728093 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 17 17:25:51.734360 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 17 17:25:51.736255 systemd-networkd[1115]: eth0: DHCPv6 lease lost Mar 17 17:25:51.748257 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 17 17:25:51.752626 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 17 17:25:51.761156 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 17 17:25:51.764913 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 17 17:25:51.768524 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 17 17:25:51.768648 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:25:51.784130 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 17 17:25:51.786617 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 17 17:25:51.786809 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 17 17:25:51.789406 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 17:25:51.790611 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:25:51.794714 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 17 17:25:51.798918 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 17 17:25:51.807197 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 17 17:25:51.807315 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:25:51.813958 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:25:51.840230 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 17 17:25:51.840460 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 17 17:25:51.845169 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 17 17:25:51.845375 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 17 17:25:51.854976 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 17 17:25:51.858028 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:25:51.864689 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 17 17:25:51.866422 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 17 17:25:51.869448 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 17 17:25:51.869532 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:25:51.872931 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 17 17:25:51.873040 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 17 17:25:51.876971 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 17 17:25:51.877066 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 17 17:25:51.879221 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 17:25:51.879315 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:25:51.906594 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 17 17:25:51.909651 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 17 17:25:51.912411 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:25:51.918236 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:25:51.918353 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:25:51.921540 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 17 17:25:51.921808 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 17 17:25:51.943799 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 17 17:25:51.946036 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 17 17:25:51.952578 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 17 17:25:51.965198 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 17 17:25:51.987781 systemd[1]: Switching root. Mar 17 17:25:52.031542 systemd-journald[251]: Journal stopped Mar 17 17:25:53.891251 systemd-journald[251]: Received SIGTERM from PID 1 (systemd). Mar 17 17:25:53.891414 kernel: SELinux: policy capability network_peer_controls=1 Mar 17 17:25:53.891464 kernel: SELinux: policy capability open_perms=1 Mar 17 17:25:53.891500 kernel: SELinux: policy capability extended_socket_class=1 Mar 17 17:25:53.891534 kernel: SELinux: policy capability always_check_network=0 Mar 17 17:25:53.891578 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 17 17:25:53.891612 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 17 17:25:53.891643 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 17 17:25:53.891673 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 17 17:25:53.891705 kernel: audit: type=1403 audit(1742232352.323:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 17 17:25:53.891790 systemd[1]: Successfully loaded SELinux policy in 49.889ms. Mar 17 17:25:53.891840 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 24.052ms. Mar 17 17:25:53.891880 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 17 17:25:53.891921 systemd[1]: Detected virtualization amazon. Mar 17 17:25:53.891955 systemd[1]: Detected architecture arm64. Mar 17 17:25:53.891986 systemd[1]: Detected first boot. Mar 17 17:25:53.892019 systemd[1]: Initializing machine ID from VM UUID. Mar 17 17:25:53.892051 zram_generator::config[1420]: No configuration found. Mar 17 17:25:53.892102 systemd[1]: Populated /etc with preset unit settings. Mar 17 17:25:53.892137 systemd[1]: Queued start job for default target multi-user.target. Mar 17 17:25:53.892170 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Mar 17 17:25:53.892204 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 17 17:25:53.892237 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 17 17:25:53.892270 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 17 17:25:53.892307 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 17 17:25:53.892340 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 17 17:25:53.892373 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 17 17:25:53.892408 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 17 17:25:53.892442 systemd[1]: Created slice user.slice - User and Session Slice. Mar 17 17:25:53.892497 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:25:53.892534 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:25:53.892565 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 17 17:25:53.892598 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 17 17:25:53.892630 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 17 17:25:53.892666 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 17 17:25:53.892698 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 17 17:25:53.892776 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:25:53.892813 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 17 17:25:53.892845 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:25:53.892880 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 17 17:25:53.892912 systemd[1]: Reached target slices.target - Slice Units. Mar 17 17:25:53.892944 systemd[1]: Reached target swap.target - Swaps. Mar 17 17:25:53.895224 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 17 17:25:53.895278 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 17 17:25:53.895323 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 17 17:25:53.895354 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 17 17:25:53.895386 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:25:53.895419 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 17 17:25:53.895452 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:25:53.895483 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 17 17:25:53.895512 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 17 17:25:53.895544 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 17 17:25:53.895575 systemd[1]: Mounting media.mount - External Media Directory... Mar 17 17:25:53.895613 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 17 17:25:53.895646 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 17 17:25:53.895680 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 17 17:25:53.895714 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 17 17:25:53.895829 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:25:53.895869 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 17 17:25:53.895906 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 17 17:25:53.895939 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:25:53.895974 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 17 17:25:53.896011 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:25:53.896041 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 17 17:25:53.896083 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:25:53.896118 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 17 17:25:53.896154 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Mar 17 17:25:53.896191 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Mar 17 17:25:53.896223 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 17 17:25:53.896254 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 17 17:25:53.896288 kernel: fuse: init (API version 7.39) Mar 17 17:25:53.896321 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 17 17:25:53.896351 kernel: loop: module loaded Mar 17 17:25:53.896380 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 17 17:25:53.896408 kernel: ACPI: bus type drm_connector registered Mar 17 17:25:53.896438 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 17 17:25:53.896468 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 17 17:25:53.896500 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 17 17:25:53.896531 systemd[1]: Mounted media.mount - External Media Directory. Mar 17 17:25:53.896567 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 17 17:25:53.896597 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 17 17:25:53.896630 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 17 17:25:53.896664 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:25:53.896700 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 17 17:25:53.896779 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 17 17:25:53.896815 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:25:53.896848 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:25:53.896880 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 17:25:53.896917 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 17 17:25:53.897012 systemd-journald[1520]: Collecting audit messages is disabled. Mar 17 17:25:53.897068 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:25:53.897103 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:25:53.897137 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 17 17:25:53.897165 systemd-journald[1520]: Journal started Mar 17 17:25:53.897214 systemd-journald[1520]: Runtime Journal (/run/log/journal/ec2786e69a82c3ce72a6a6a1720d632d) is 8.0M, max 75.3M, 67.3M free. Mar 17 17:25:53.900177 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 17 17:25:53.904812 systemd[1]: Started systemd-journald.service - Journal Service. Mar 17 17:25:53.911811 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:25:53.912261 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:25:53.918076 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 17 17:25:53.922681 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 17 17:25:53.926322 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 17 17:25:53.931601 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 17 17:25:53.962371 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 17 17:25:53.974092 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 17 17:25:53.985918 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 17 17:25:53.991045 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 17 17:25:54.007092 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 17 17:25:54.021384 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 17 17:25:54.026967 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 17:25:54.038046 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 17 17:25:54.040330 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 17 17:25:54.055161 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 17 17:25:54.078089 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 17 17:25:54.086637 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 17 17:25:54.105213 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 17 17:25:54.121609 systemd-journald[1520]: Time spent on flushing to /var/log/journal/ec2786e69a82c3ce72a6a6a1720d632d is 90.704ms for 877 entries. Mar 17 17:25:54.121609 systemd-journald[1520]: System Journal (/var/log/journal/ec2786e69a82c3ce72a6a6a1720d632d) is 8.0M, max 195.6M, 187.6M free. Mar 17 17:25:54.226514 systemd-journald[1520]: Received client request to flush runtime journal. Mar 17 17:25:54.140646 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 17 17:25:54.144118 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 17 17:25:54.194540 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:25:54.234871 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 17 17:25:54.246559 systemd-tmpfiles[1572]: ACLs are not supported, ignoring. Mar 17 17:25:54.247204 systemd-tmpfiles[1572]: ACLs are not supported, ignoring. Mar 17 17:25:54.257332 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 17 17:25:54.264296 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:25:54.283091 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 17 17:25:54.295027 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 17 17:25:54.349078 udevadm[1590]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 17 17:25:54.384887 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 17 17:25:54.395104 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 17 17:25:54.444501 systemd-tmpfiles[1594]: ACLs are not supported, ignoring. Mar 17 17:25:54.445242 systemd-tmpfiles[1594]: ACLs are not supported, ignoring. Mar 17 17:25:54.456968 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:25:55.151506 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 17 17:25:55.161087 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:25:55.226635 systemd-udevd[1600]: Using default interface naming scheme 'v255'. Mar 17 17:25:55.283466 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:25:55.299099 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 17 17:25:55.344793 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 17 17:25:55.456905 (udev-worker)[1605]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:25:55.463083 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Mar 17 17:25:55.532156 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 17 17:25:55.708496 systemd-networkd[1604]: lo: Link UP Mar 17 17:25:55.708516 systemd-networkd[1604]: lo: Gained carrier Mar 17 17:25:55.712574 systemd-networkd[1604]: Enumeration completed Mar 17 17:25:55.713392 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 17 17:25:55.716171 systemd-networkd[1604]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:25:55.716179 systemd-networkd[1604]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 17:25:55.722031 systemd-networkd[1604]: eth0: Link UP Mar 17 17:25:55.722467 systemd-networkd[1604]: eth0: Gained carrier Mar 17 17:25:55.722520 systemd-networkd[1604]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:25:55.731210 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 17 17:25:55.742619 systemd-networkd[1604]: eth0: DHCPv4 address 172.31.31.32/20, gateway 172.31.16.1 acquired from 172.31.16.1 Mar 17 17:25:55.820837 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (1606) Mar 17 17:25:55.861947 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:25:56.050019 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 17 17:25:56.070697 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:25:56.100709 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Mar 17 17:25:56.111068 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 17 17:25:56.147825 lvm[1729]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 17:25:56.187713 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 17 17:25:56.191724 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:25:56.204137 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 17 17:25:56.222345 lvm[1732]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 17:25:56.267117 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 17 17:25:56.270419 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 17 17:25:56.275300 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 17 17:25:56.275524 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 17 17:25:56.277904 systemd[1]: Reached target machines.target - Containers. Mar 17 17:25:56.282354 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 17 17:25:56.290073 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 17 17:25:56.302198 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 17 17:25:56.305126 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:25:56.310220 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 17 17:25:56.318421 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 17 17:25:56.335043 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 17 17:25:56.342931 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 17 17:25:56.388519 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 17 17:25:56.408870 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 17 17:25:56.410542 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 17 17:25:56.413873 kernel: loop0: detected capacity change from 0 to 53784 Mar 17 17:25:56.513909 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 17 17:25:56.537803 kernel: loop1: detected capacity change from 0 to 116808 Mar 17 17:25:56.591904 kernel: loop2: detected capacity change from 0 to 113536 Mar 17 17:25:56.652261 kernel: loop3: detected capacity change from 0 to 194096 Mar 17 17:25:56.774863 kernel: loop4: detected capacity change from 0 to 53784 Mar 17 17:25:56.787788 kernel: loop5: detected capacity change from 0 to 116808 Mar 17 17:25:56.807789 kernel: loop6: detected capacity change from 0 to 113536 Mar 17 17:25:56.836254 kernel: loop7: detected capacity change from 0 to 194096 Mar 17 17:25:56.878500 (sd-merge)[1754]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Mar 17 17:25:56.879598 (sd-merge)[1754]: Merged extensions into '/usr'. Mar 17 17:25:56.890128 systemd[1]: Reloading requested from client PID 1740 ('systemd-sysext') (unit systemd-sysext.service)... Mar 17 17:25:56.890905 systemd[1]: Reloading... Mar 17 17:25:57.043787 zram_generator::config[1783]: No configuration found. Mar 17 17:25:57.069708 ldconfig[1736]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 17 17:25:57.340523 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:25:57.360935 systemd-networkd[1604]: eth0: Gained IPv6LL Mar 17 17:25:57.491136 systemd[1]: Reloading finished in 599 ms. Mar 17 17:25:57.519469 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 17 17:25:57.523415 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 17 17:25:57.528084 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 17 17:25:57.554137 systemd[1]: Starting ensure-sysext.service... Mar 17 17:25:57.561079 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 17 17:25:57.570056 systemd[1]: Reloading requested from client PID 1844 ('systemctl') (unit ensure-sysext.service)... Mar 17 17:25:57.570252 systemd[1]: Reloading... Mar 17 17:25:57.631183 systemd-tmpfiles[1845]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 17 17:25:57.632032 systemd-tmpfiles[1845]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 17 17:25:57.634127 systemd-tmpfiles[1845]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 17 17:25:57.634834 systemd-tmpfiles[1845]: ACLs are not supported, ignoring. Mar 17 17:25:57.635030 systemd-tmpfiles[1845]: ACLs are not supported, ignoring. Mar 17 17:25:57.641355 systemd-tmpfiles[1845]: Detected autofs mount point /boot during canonicalization of boot. Mar 17 17:25:57.641385 systemd-tmpfiles[1845]: Skipping /boot Mar 17 17:25:57.667240 systemd-tmpfiles[1845]: Detected autofs mount point /boot during canonicalization of boot. Mar 17 17:25:57.667280 systemd-tmpfiles[1845]: Skipping /boot Mar 17 17:25:57.756329 zram_generator::config[1872]: No configuration found. Mar 17 17:25:58.019311 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:25:58.165172 systemd[1]: Reloading finished in 594 ms. Mar 17 17:25:58.203835 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:25:58.224106 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 17 17:25:58.244099 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 17 17:25:58.253083 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 17 17:25:58.264029 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 17 17:25:58.283904 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 17 17:25:58.303075 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:25:58.315041 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:25:58.333403 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:25:58.360032 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:25:58.362502 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:25:58.365140 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:25:58.365604 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:25:58.377823 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:25:58.378207 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:25:58.408662 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:25:58.410228 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:25:58.429518 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 17 17:25:58.436026 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:25:58.446812 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:25:58.468699 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:25:58.476447 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:25:58.479600 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:25:58.483084 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 17 17:25:58.495967 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:25:58.507130 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 17 17:25:58.509378 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:25:58.509879 systemd[1]: Reached target time-set.target - System Time Set. Mar 17 17:25:58.526235 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 17 17:25:58.535375 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:25:58.537909 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:25:58.544469 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:25:58.549259 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:25:58.559161 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:25:58.559513 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:25:58.563099 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 17:25:58.564519 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 17 17:25:58.587257 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 17:25:58.591260 augenrules[1979]: No rules Mar 17 17:25:58.587594 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 17 17:25:58.595538 systemd[1]: audit-rules.service: Deactivated successfully. Mar 17 17:25:58.599406 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 17 17:25:58.608231 systemd[1]: Finished ensure-sysext.service. Mar 17 17:25:58.627037 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 17 17:25:58.633444 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 17:25:58.640462 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 17 17:25:58.672548 systemd-resolved[1935]: Positive Trust Anchors: Mar 17 17:25:58.672622 systemd-resolved[1935]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 17:25:58.672684 systemd-resolved[1935]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 17 17:25:58.680435 systemd-resolved[1935]: Defaulting to hostname 'linux'. Mar 17 17:25:58.683712 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 17 17:25:58.686103 systemd[1]: Reached target network.target - Network. Mar 17 17:25:58.687931 systemd[1]: Reached target network-online.target - Network is Online. Mar 17 17:25:58.690046 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:25:58.692310 systemd[1]: Reached target sysinit.target - System Initialization. Mar 17 17:25:58.694474 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 17 17:25:58.696866 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 17 17:25:58.699507 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 17 17:25:58.701852 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 17 17:25:58.704229 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 17 17:25:58.706594 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 17 17:25:58.706654 systemd[1]: Reached target paths.target - Path Units. Mar 17 17:25:58.708322 systemd[1]: Reached target timers.target - Timer Units. Mar 17 17:25:58.711971 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 17 17:25:58.717208 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 17 17:25:58.721278 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 17 17:25:58.725694 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 17 17:25:58.727925 systemd[1]: Reached target sockets.target - Socket Units. Mar 17 17:25:58.730946 systemd[1]: Reached target basic.target - Basic System. Mar 17 17:25:58.733351 systemd[1]: System is tainted: cgroupsv1 Mar 17 17:25:58.733450 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 17 17:25:58.733500 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 17 17:25:58.743833 systemd[1]: Starting containerd.service - containerd container runtime... Mar 17 17:25:58.753058 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 17 17:25:58.764027 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 17 17:25:58.772937 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 17 17:25:58.784055 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 17 17:25:58.789879 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 17 17:25:58.808706 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:25:58.822900 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 17 17:25:58.826365 jq[2004]: false Mar 17 17:25:58.848002 systemd[1]: Started ntpd.service - Network Time Service. Mar 17 17:25:58.886036 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 17 17:25:58.889851 extend-filesystems[2005]: Found loop4 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found loop5 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found loop6 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found loop7 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p1 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p2 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p3 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found usr Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p4 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p6 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p7 Mar 17 17:25:58.893326 extend-filesystems[2005]: Found nvme0n1p9 Mar 17 17:25:58.893326 extend-filesystems[2005]: Checking size of /dev/nvme0n1p9 Mar 17 17:25:58.897690 dbus-daemon[2003]: [system] SELinux support is enabled Mar 17 17:25:58.936172 systemd[1]: Starting setup-oem.service - Setup OEM... Mar 17 17:25:58.992813 extend-filesystems[2005]: Resized partition /dev/nvme0n1p9 Mar 17 17:25:58.908305 dbus-daemon[2003]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1604 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Mar 17 17:25:58.996714 extend-filesystems[2022]: resize2fs 1.47.1 (20-May-2024) Mar 17 17:25:59.026040 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Mar 17 17:25:58.997163 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 17 17:25:59.016061 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 17 17:25:59.037145 ntpd[2011]: ntpd 4.2.8p17@1.4004-o Mon Mar 17 15:34:53 UTC 2025 (1): Starting Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: ntpd 4.2.8p17@1.4004-o Mon Mar 17 15:34:53 UTC 2025 (1): Starting Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: ---------------------------------------------------- Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: ntp-4 is maintained by Network Time Foundation, Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: corporation. Support and training for ntp-4 are Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: available at https://www.nwtime.org/support Mar 17 17:25:59.045905 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: ---------------------------------------------------- Mar 17 17:25:59.037221 ntpd[2011]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Mar 17 17:25:59.047545 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 17 17:25:59.037264 ntpd[2011]: ---------------------------------------------------- Mar 17 17:25:59.037287 ntpd[2011]: ntp-4 is maintained by Network Time Foundation, Mar 17 17:25:59.037306 ntpd[2011]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Mar 17 17:25:59.054811 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 17 17:25:59.066143 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: proto: precision = 0.096 usec (-23) Mar 17 17:25:59.066143 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: basedate set to 2025-03-05 Mar 17 17:25:59.066143 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: gps base set to 2025-03-09 (week 2357) Mar 17 17:25:59.037325 ntpd[2011]: corporation. Support and training for ntp-4 are Mar 17 17:25:59.037344 ntpd[2011]: available at https://www.nwtime.org/support Mar 17 17:25:59.037362 ntpd[2011]: ---------------------------------------------------- Mar 17 17:25:59.056318 ntpd[2011]: proto: precision = 0.096 usec (-23) Mar 17 17:25:59.056844 ntpd[2011]: basedate set to 2025-03-05 Mar 17 17:25:59.056874 ntpd[2011]: gps base set to 2025-03-09 (week 2357) Mar 17 17:25:59.070113 systemd[1]: Starting update-engine.service - Update Engine... Mar 17 17:25:59.092009 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 17 17:25:59.089074 ntpd[2011]: Listen and drop on 0 v6wildcard [::]:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen and drop on 0 v6wildcard [::]:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen normally on 2 lo 127.0.0.1:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen normally on 3 eth0 172.31.31.32:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen normally on 4 lo [::1]:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listen normally on 5 eth0 [fe80::46c:a0ff:fe0a:3499%2]:123 Mar 17 17:25:59.093257 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: Listening on routing socket on fd #22 for interface updates Mar 17 17:25:59.089150 ntpd[2011]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Mar 17 17:25:59.089441 ntpd[2011]: Listen normally on 2 lo 127.0.0.1:123 Mar 17 17:25:59.101428 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 17 17:25:59.089510 ntpd[2011]: Listen normally on 3 eth0 172.31.31.32:123 Mar 17 17:25:59.089580 ntpd[2011]: Listen normally on 4 lo [::1]:123 Mar 17 17:25:59.089656 ntpd[2011]: Listen normally on 5 eth0 [fe80::46c:a0ff:fe0a:3499%2]:123 Mar 17 17:25:59.115012 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Mar 17 17:25:59.115012 ntpd[2011]: 17 Mar 17:25:59 ntpd[2011]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Mar 17 17:25:59.089717 ntpd[2011]: Listening on routing socket on fd #22 for interface updates Mar 17 17:25:59.114858 ntpd[2011]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Mar 17 17:25:59.114909 ntpd[2011]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Mar 17 17:25:59.124444 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 17 17:25:59.126100 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 17 17:25:59.137618 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 17 17:25:59.141484 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 17 17:25:59.152630 update_engine[2032]: I20250317 17:25:59.151921 2032 main.cc:92] Flatcar Update Engine starting Mar 17 17:25:59.159808 update_engine[2032]: I20250317 17:25:59.157518 2032 update_check_scheduler.cc:74] Next update check in 11m4s Mar 17 17:25:59.184822 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Mar 17 17:25:59.211310 jq[2038]: true Mar 17 17:25:59.244693 systemd[1]: motdgen.service: Deactivated successfully. Mar 17 17:25:59.245330 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 17 17:25:59.260936 dbus-daemon[2003]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 17 17:25:59.264574 extend-filesystems[2022]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Mar 17 17:25:59.264574 extend-filesystems[2022]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 17 17:25:59.264574 extend-filesystems[2022]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Mar 17 17:25:59.284124 extend-filesystems[2005]: Resized filesystem in /dev/nvme0n1p9 Mar 17 17:25:59.290190 coreos-metadata[2002]: Mar 17 17:25:59.289 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Mar 17 17:25:59.297647 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 17 17:25:59.298244 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 17 17:25:59.312761 coreos-metadata[2002]: Mar 17 17:25:59.310 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Mar 17 17:25:59.312761 coreos-metadata[2002]: Mar 17 17:25:59.310 INFO Fetch successful Mar 17 17:25:59.312761 coreos-metadata[2002]: Mar 17 17:25:59.310 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.321 INFO Fetch successful Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.321 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.321 INFO Fetch successful Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.321 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.322 INFO Fetch successful Mar 17 17:25:59.322942 coreos-metadata[2002]: Mar 17 17:25:59.322 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Mar 17 17:25:59.330881 coreos-metadata[2002]: Mar 17 17:25:59.328 INFO Fetch failed with 404: resource not found Mar 17 17:25:59.330881 coreos-metadata[2002]: Mar 17 17:25:59.328 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Mar 17 17:25:59.332129 coreos-metadata[2002]: Mar 17 17:25:59.332 INFO Fetch successful Mar 17 17:25:59.332129 coreos-metadata[2002]: Mar 17 17:25:59.332 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Mar 17 17:25:59.334398 (ntainerd)[2072]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 17 17:25:59.338298 coreos-metadata[2002]: Mar 17 17:25:59.338 INFO Fetch successful Mar 17 17:25:59.338298 coreos-metadata[2002]: Mar 17 17:25:59.338 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Mar 17 17:25:59.339106 jq[2057]: true Mar 17 17:25:59.359717 coreos-metadata[2002]: Mar 17 17:25:59.359 INFO Fetch successful Mar 17 17:25:59.362624 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 17 17:25:59.370785 coreos-metadata[2002]: Mar 17 17:25:59.366 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Mar 17 17:25:59.375484 coreos-metadata[2002]: Mar 17 17:25:59.373 INFO Fetch successful Mar 17 17:25:59.375484 coreos-metadata[2002]: Mar 17 17:25:59.375 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Mar 17 17:25:59.379516 coreos-metadata[2002]: Mar 17 17:25:59.376 INFO Fetch successful Mar 17 17:25:59.383305 systemd[1]: Finished setup-oem.service - Setup OEM. Mar 17 17:25:59.430376 systemd[1]: Started update-engine.service - Update Engine. Mar 17 17:25:59.447098 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Mar 17 17:25:59.451807 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 17 17:25:59.451876 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 17 17:25:59.471051 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Mar 17 17:25:59.475235 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 17 17:25:59.475277 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 17 17:25:59.482249 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 17 17:25:59.495304 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 17 17:25:59.577776 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (2086) Mar 17 17:25:59.639546 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 17 17:25:59.645115 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 17 17:25:59.653073 bash[2121]: Updated "/home/core/.ssh/authorized_keys" Mar 17 17:25:59.659778 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 17 17:25:59.673029 systemd[1]: Starting sshkeys.service... Mar 17 17:25:59.739553 amazon-ssm-agent[2100]: Initializing new seelog logger Mar 17 17:25:59.744851 systemd-logind[2030]: Watching system buttons on /dev/input/event0 (Power Button) Mar 17 17:25:59.746812 amazon-ssm-agent[2100]: New Seelog Logger Creation Complete Mar 17 17:25:59.746812 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.746812 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.746812 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 processing appconfig overrides Mar 17 17:25:59.744895 systemd-logind[2030]: Watching system buttons on /dev/input/event1 (Sleep Button) Mar 17 17:25:59.747267 systemd-logind[2030]: New seat seat0. Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 processing appconfig overrides Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 processing appconfig overrides Mar 17 17:25:59.756850 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO Proxy environment variables: Mar 17 17:25:59.764221 systemd[1]: Started systemd-logind.service - User Login Management. Mar 17 17:25:59.774650 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 17 17:25:59.787644 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.787644 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Mar 17 17:25:59.787644 amazon-ssm-agent[2100]: 2025/03/17 17:25:59 processing appconfig overrides Mar 17 17:25:59.785816 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 17 17:25:59.859412 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO http_proxy: Mar 17 17:25:59.920125 locksmithd[2106]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 17 17:25:59.959050 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO no_proxy: Mar 17 17:26:00.049010 coreos-metadata[2144]: Mar 17 17:26:00.048 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Mar 17 17:26:00.050397 coreos-metadata[2144]: Mar 17 17:26:00.050 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Mar 17 17:26:00.054878 coreos-metadata[2144]: Mar 17 17:26:00.054 INFO Fetch successful Mar 17 17:26:00.055870 coreos-metadata[2144]: Mar 17 17:26:00.055 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Mar 17 17:26:00.058222 coreos-metadata[2144]: Mar 17 17:26:00.058 INFO Fetch successful Mar 17 17:26:00.065656 unknown[2144]: wrote ssh authorized keys file for user: core Mar 17 17:26:00.075031 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO https_proxy: Mar 17 17:26:00.145886 update-ssh-keys[2202]: Updated "/home/core/.ssh/authorized_keys" Mar 17 17:26:00.149953 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 17 17:26:00.177425 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO Checking if agent identity type OnPrem can be assumed Mar 17 17:26:00.177679 systemd[1]: Finished sshkeys.service. Mar 17 17:26:00.288881 amazon-ssm-agent[2100]: 2025-03-17 17:25:59 INFO Checking if agent identity type EC2 can be assumed Mar 17 17:26:00.305832 containerd[2072]: time="2025-03-17T17:26:00.303351946Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Mar 17 17:26:00.387178 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO Agent will take identity from EC2 Mar 17 17:26:00.486140 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] using named pipe channel for IPC Mar 17 17:26:00.538773 containerd[2072]: time="2025-03-17T17:26:00.537903083Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.545851 containerd[2072]: time="2025-03-17T17:26:00.545759783Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.83-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:26:00.545851 containerd[2072]: time="2025-03-17T17:26:00.545837147Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 17 17:26:00.546015 containerd[2072]: time="2025-03-17T17:26:00.545878703Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 17 17:26:00.546298 containerd[2072]: time="2025-03-17T17:26:00.546218159Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 17 17:26:00.546391 containerd[2072]: time="2025-03-17T17:26:00.546295319Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.546517 containerd[2072]: time="2025-03-17T17:26:00.546460307Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:26:00.546517 containerd[2072]: time="2025-03-17T17:26:00.546509099Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.551123 containerd[2072]: time="2025-03-17T17:26:00.551050499Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:26:00.551123 containerd[2072]: time="2025-03-17T17:26:00.551114327Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.551322 containerd[2072]: time="2025-03-17T17:26:00.551154023Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:26:00.551322 containerd[2072]: time="2025-03-17T17:26:00.551178815Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.551471 containerd[2072]: time="2025-03-17T17:26:00.551415047Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.552023 containerd[2072]: time="2025-03-17T17:26:00.551952731Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:26:00.552374 containerd[2072]: time="2025-03-17T17:26:00.552310079Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:26:00.552374 containerd[2072]: time="2025-03-17T17:26:00.552364991Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 17 17:26:00.552651 containerd[2072]: time="2025-03-17T17:26:00.552600635Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 17 17:26:00.556465 dbus-daemon[2003]: [system] Successfully activated service 'org.freedesktop.hostname1' Mar 17 17:26:00.556808 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Mar 17 17:26:00.561138 containerd[2072]: time="2025-03-17T17:26:00.561045263Z" level=info msg="metadata content store policy set" policy=shared Mar 17 17:26:00.564347 dbus-daemon[2003]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=2104 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Mar 17 17:26:00.566709 containerd[2072]: time="2025-03-17T17:26:00.566643431Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 17 17:26:00.566832 containerd[2072]: time="2025-03-17T17:26:00.566766683Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 17 17:26:00.566832 containerd[2072]: time="2025-03-17T17:26:00.566803427Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 17 17:26:00.566949 containerd[2072]: time="2025-03-17T17:26:00.566841803Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 17 17:26:00.566949 containerd[2072]: time="2025-03-17T17:26:00.566876867Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 17 17:26:00.567174 containerd[2072]: time="2025-03-17T17:26:00.567129947Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 17 17:26:00.582992 systemd[1]: Starting polkit.service - Authorization Manager... Mar 17 17:26:00.585356 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] using named pipe channel for IPC Mar 17 17:26:00.592035 containerd[2072]: time="2025-03-17T17:26:00.589992059Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 17 17:26:00.599807 containerd[2072]: time="2025-03-17T17:26:00.596696711Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 17 17:26:00.599949 containerd[2072]: time="2025-03-17T17:26:00.599871707Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 17 17:26:00.600001 containerd[2072]: time="2025-03-17T17:26:00.599960939Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 17 17:26:00.600218 containerd[2072]: time="2025-03-17T17:26:00.600171479Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.600830 containerd[2072]: time="2025-03-17T17:26:00.600781559Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.600899 containerd[2072]: time="2025-03-17T17:26:00.600868787Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.600982 containerd[2072]: time="2025-03-17T17:26:00.600944351Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.601064 containerd[2072]: time="2025-03-17T17:26:00.601023551Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.604773 containerd[2072]: time="2025-03-17T17:26:00.601074899Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.607763 containerd[2072]: time="2025-03-17T17:26:00.604552139Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.607763 containerd[2072]: time="2025-03-17T17:26:00.606984936Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 17 17:26:00.607763 containerd[2072]: time="2025-03-17T17:26:00.607095828Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.608883 containerd[2072]: time="2025-03-17T17:26:00.608800956Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.608989 containerd[2072]: time="2025-03-17T17:26:00.608903172Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.608989 containerd[2072]: time="2025-03-17T17:26:00.608976744Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.609178 containerd[2072]: time="2025-03-17T17:26:00.609073572Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617370 containerd[2072]: time="2025-03-17T17:26:00.617304360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617515 containerd[2072]: time="2025-03-17T17:26:00.617375232Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617515 containerd[2072]: time="2025-03-17T17:26:00.617411292Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617515 containerd[2072]: time="2025-03-17T17:26:00.617444592Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617515 containerd[2072]: time="2025-03-17T17:26:00.617485044Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617681 containerd[2072]: time="2025-03-17T17:26:00.617515464Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617681 containerd[2072]: time="2025-03-17T17:26:00.617548464Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617681 containerd[2072]: time="2025-03-17T17:26:00.617580024Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617681 containerd[2072]: time="2025-03-17T17:26:00.617614260Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 17 17:26:00.617681 containerd[2072]: time="2025-03-17T17:26:00.617669100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617975 containerd[2072]: time="2025-03-17T17:26:00.617703960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.617975 containerd[2072]: time="2025-03-17T17:26:00.617767968Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 17 17:26:00.617975 containerd[2072]: time="2025-03-17T17:26:00.617913912Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 17 17:26:00.617975 containerd[2072]: time="2025-03-17T17:26:00.617957100Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.617982948Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.618013128Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.618036384Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.618065568Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.618088980Z" level=info msg="NRI interface is disabled by configuration." Mar 17 17:26:00.618149 containerd[2072]: time="2025-03-17T17:26:00.618113424Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 17 17:26:00.624814 containerd[2072]: time="2025-03-17T17:26:00.622347684Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 17 17:26:00.624814 containerd[2072]: time="2025-03-17T17:26:00.622472868Z" level=info msg="Connect containerd service" Mar 17 17:26:00.624814 containerd[2072]: time="2025-03-17T17:26:00.622557684Z" level=info msg="using legacy CRI server" Mar 17 17:26:00.624814 containerd[2072]: time="2025-03-17T17:26:00.622576440Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 17 17:26:00.624814 containerd[2072]: time="2025-03-17T17:26:00.622853988Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 17 17:26:00.628440 containerd[2072]: time="2025-03-17T17:26:00.628104036Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 17:26:00.632593 containerd[2072]: time="2025-03-17T17:26:00.630463140Z" level=info msg="Start subscribing containerd event" Mar 17 17:26:00.632851 containerd[2072]: time="2025-03-17T17:26:00.632562012Z" level=info msg="Start recovering state" Mar 17 17:26:00.633653 containerd[2072]: time="2025-03-17T17:26:00.633258216Z" level=info msg="Start event monitor" Mar 17 17:26:00.633653 containerd[2072]: time="2025-03-17T17:26:00.633445476Z" level=info msg="Start snapshots syncer" Mar 17 17:26:00.633653 containerd[2072]: time="2025-03-17T17:26:00.633500364Z" level=info msg="Start cni network conf syncer for default" Mar 17 17:26:00.633653 containerd[2072]: time="2025-03-17T17:26:00.633524004Z" level=info msg="Start streaming server" Mar 17 17:26:00.634816 containerd[2072]: time="2025-03-17T17:26:00.634457904Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 17 17:26:00.635002 containerd[2072]: time="2025-03-17T17:26:00.634965768Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 17 17:26:00.639498 systemd[1]: Started containerd.service - containerd container runtime. Mar 17 17:26:00.641779 containerd[2072]: time="2025-03-17T17:26:00.639444240Z" level=info msg="containerd successfully booted in 0.344914s" Mar 17 17:26:00.652814 polkitd[2247]: Started polkitd version 121 Mar 17 17:26:00.669894 polkitd[2247]: Loading rules from directory /etc/polkit-1/rules.d Mar 17 17:26:00.670029 polkitd[2247]: Loading rules from directory /usr/share/polkit-1/rules.d Mar 17 17:26:00.671270 polkitd[2247]: Finished loading, compiling and executing 2 rules Mar 17 17:26:00.672773 dbus-daemon[2003]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Mar 17 17:26:00.673042 systemd[1]: Started polkit.service - Authorization Manager. Mar 17 17:26:00.679862 polkitd[2247]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Mar 17 17:26:00.684620 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] using named pipe channel for IPC Mar 17 17:26:00.715476 systemd-hostnamed[2104]: Hostname set to (transient) Mar 17 17:26:00.716122 systemd-resolved[1935]: System hostname changed to 'ip-172-31-31-32'. Mar 17 17:26:00.783915 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Mar 17 17:26:00.887862 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 Mar 17 17:26:00.988151 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] Starting Core Agent Mar 17 17:26:01.088872 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [amazon-ssm-agent] registrar detected. Attempting registration Mar 17 17:26:01.192618 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [Registrar] Starting registrar module Mar 17 17:26:01.292842 amazon-ssm-agent[2100]: 2025-03-17 17:26:00 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Mar 17 17:26:01.754491 amazon-ssm-agent[2100]: 2025-03-17 17:26:01 INFO [EC2Identity] EC2 registration was successful. Mar 17 17:26:01.784033 amazon-ssm-agent[2100]: 2025-03-17 17:26:01 INFO [CredentialRefresher] credentialRefresher has started Mar 17 17:26:01.786364 amazon-ssm-agent[2100]: 2025-03-17 17:26:01 INFO [CredentialRefresher] Starting credentials refresher loop Mar 17 17:26:01.786364 amazon-ssm-agent[2100]: 2025-03-17 17:26:01 INFO EC2RoleProvider Successfully connected with instance profile role credentials Mar 17 17:26:01.855323 amazon-ssm-agent[2100]: 2025-03-17 17:26:01 INFO [CredentialRefresher] Next credential rotation will be in 30.024957046 minutes Mar 17 17:26:02.087024 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:26:02.090067 (kubelet)[2267]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 17 17:26:02.741896 sshd_keygen[2049]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 17 17:26:02.783599 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 17 17:26:02.797283 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 17 17:26:02.820451 amazon-ssm-agent[2100]: 2025-03-17 17:26:02 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Mar 17 17:26:02.840315 systemd[1]: issuegen.service: Deactivated successfully. Mar 17 17:26:02.840917 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 17 17:26:02.852301 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 17 17:26:02.894607 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 17 17:26:02.912049 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 17 17:26:02.923409 amazon-ssm-agent[2100]: 2025-03-17 17:26:02 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2283) started Mar 17 17:26:02.929434 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 17 17:26:02.932397 systemd[1]: Reached target getty.target - Login Prompts. Mar 17 17:26:02.935565 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 17 17:26:02.938892 systemd[1]: Startup finished in 8.707s (kernel) + 10.662s (userspace) = 19.370s. Mar 17 17:26:03.023894 amazon-ssm-agent[2100]: 2025-03-17 17:26:02 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Mar 17 17:26:03.526859 kubelet[2267]: E0317 17:26:03.526720 2267 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 17:26:03.531995 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 17:26:03.532408 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 17:26:06.970366 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 17 17:26:06.982154 systemd[1]: Started sshd@0-172.31.31.32:22-139.178.68.195:58712.service - OpenSSH per-connection server daemon (139.178.68.195:58712). Mar 17 17:26:07.182038 sshd[2310]: Accepted publickey for core from 139.178.68.195 port 58712 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:07.184782 sshd-session[2310]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:07.199780 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 17 17:26:07.211150 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 17 17:26:07.215295 systemd-logind[2030]: New session 1 of user core. Mar 17 17:26:07.235646 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 17 17:26:07.248402 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 17 17:26:07.258671 (systemd)[2316]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 17 17:26:07.464099 systemd[2316]: Queued start job for default target default.target. Mar 17 17:26:07.465245 systemd[2316]: Created slice app.slice - User Application Slice. Mar 17 17:26:07.465290 systemd[2316]: Reached target paths.target - Paths. Mar 17 17:26:07.465321 systemd[2316]: Reached target timers.target - Timers. Mar 17 17:26:07.471905 systemd[2316]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 17 17:26:07.499136 systemd[2316]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 17 17:26:07.499263 systemd[2316]: Reached target sockets.target - Sockets. Mar 17 17:26:07.499295 systemd[2316]: Reached target basic.target - Basic System. Mar 17 17:26:07.499377 systemd[2316]: Reached target default.target - Main User Target. Mar 17 17:26:07.499437 systemd[2316]: Startup finished in 229ms. Mar 17 17:26:07.499968 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 17 17:26:07.509325 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 17 17:26:07.664342 systemd[1]: Started sshd@1-172.31.31.32:22-139.178.68.195:58728.service - OpenSSH per-connection server daemon (139.178.68.195:58728). Mar 17 17:26:07.860780 sshd[2328]: Accepted publickey for core from 139.178.68.195 port 58728 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:07.863027 sshd-session[2328]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:07.871763 systemd-logind[2030]: New session 2 of user core. Mar 17 17:26:07.879215 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 17 17:26:08.009048 sshd[2331]: Connection closed by 139.178.68.195 port 58728 Mar 17 17:26:08.009864 sshd-session[2328]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:08.016827 systemd-logind[2030]: Session 2 logged out. Waiting for processes to exit. Mar 17 17:26:08.018077 systemd[1]: sshd@1-172.31.31.32:22-139.178.68.195:58728.service: Deactivated successfully. Mar 17 17:26:08.022519 systemd[1]: session-2.scope: Deactivated successfully. Mar 17 17:26:08.025283 systemd-logind[2030]: Removed session 2. Mar 17 17:26:08.041247 systemd[1]: Started sshd@2-172.31.31.32:22-139.178.68.195:58742.service - OpenSSH per-connection server daemon (139.178.68.195:58742). Mar 17 17:26:08.235702 sshd[2336]: Accepted publickey for core from 139.178.68.195 port 58742 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:08.238613 sshd-session[2336]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:08.247703 systemd-logind[2030]: New session 3 of user core. Mar 17 17:26:08.255523 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 17 17:26:08.378183 sshd[2339]: Connection closed by 139.178.68.195 port 58742 Mar 17 17:26:08.377475 sshd-session[2336]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:08.384096 systemd[1]: sshd@2-172.31.31.32:22-139.178.68.195:58742.service: Deactivated successfully. Mar 17 17:26:08.389030 systemd[1]: session-3.scope: Deactivated successfully. Mar 17 17:26:08.389033 systemd-logind[2030]: Session 3 logged out. Waiting for processes to exit. Mar 17 17:26:08.391901 systemd-logind[2030]: Removed session 3. Mar 17 17:26:08.412158 systemd[1]: Started sshd@3-172.31.31.32:22-139.178.68.195:58750.service - OpenSSH per-connection server daemon (139.178.68.195:58750). Mar 17 17:26:08.593339 sshd[2344]: Accepted publickey for core from 139.178.68.195 port 58750 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:08.595136 sshd-session[2344]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:08.602468 systemd-logind[2030]: New session 4 of user core. Mar 17 17:26:08.614367 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 17 17:26:08.744799 sshd[2347]: Connection closed by 139.178.68.195 port 58750 Mar 17 17:26:08.744592 sshd-session[2344]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:08.752511 systemd[1]: sshd@3-172.31.31.32:22-139.178.68.195:58750.service: Deactivated successfully. Mar 17 17:26:08.753954 systemd-logind[2030]: Session 4 logged out. Waiting for processes to exit. Mar 17 17:26:08.758073 systemd[1]: session-4.scope: Deactivated successfully. Mar 17 17:26:08.759811 systemd-logind[2030]: Removed session 4. Mar 17 17:26:08.780176 systemd[1]: Started sshd@4-172.31.31.32:22-139.178.68.195:58766.service - OpenSSH per-connection server daemon (139.178.68.195:58766). Mar 17 17:26:08.966183 sshd[2352]: Accepted publickey for core from 139.178.68.195 port 58766 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:08.969156 sshd-session[2352]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:08.977417 systemd-logind[2030]: New session 5 of user core. Mar 17 17:26:08.988355 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 17 17:26:09.107487 sudo[2356]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 17 17:26:09.108158 sudo[2356]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:26:09.126814 sudo[2356]: pam_unix(sudo:session): session closed for user root Mar 17 17:26:09.150789 sshd[2355]: Connection closed by 139.178.68.195 port 58766 Mar 17 17:26:09.151884 sshd-session[2352]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:09.157425 systemd[1]: sshd@4-172.31.31.32:22-139.178.68.195:58766.service: Deactivated successfully. Mar 17 17:26:09.164553 systemd-logind[2030]: Session 5 logged out. Waiting for processes to exit. Mar 17 17:26:09.165478 systemd[1]: session-5.scope: Deactivated successfully. Mar 17 17:26:09.167530 systemd-logind[2030]: Removed session 5. Mar 17 17:26:09.186180 systemd[1]: Started sshd@5-172.31.31.32:22-139.178.68.195:58782.service - OpenSSH per-connection server daemon (139.178.68.195:58782). Mar 17 17:26:09.368695 sshd[2361]: Accepted publickey for core from 139.178.68.195 port 58782 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:09.371921 sshd-session[2361]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:09.380081 systemd-logind[2030]: New session 6 of user core. Mar 17 17:26:09.391317 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 17 17:26:09.497590 sudo[2366]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 17 17:26:09.498799 sudo[2366]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:26:09.505186 sudo[2366]: pam_unix(sudo:session): session closed for user root Mar 17 17:26:09.515102 sudo[2365]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 17 17:26:09.515745 sudo[2365]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:26:09.542374 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 17 17:26:09.587538 augenrules[2388]: No rules Mar 17 17:26:09.589565 systemd[1]: audit-rules.service: Deactivated successfully. Mar 17 17:26:09.591207 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 17 17:26:09.594784 sudo[2365]: pam_unix(sudo:session): session closed for user root Mar 17 17:26:09.617959 sshd[2364]: Connection closed by 139.178.68.195 port 58782 Mar 17 17:26:09.618900 sshd-session[2361]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:09.626523 systemd[1]: sshd@5-172.31.31.32:22-139.178.68.195:58782.service: Deactivated successfully. Mar 17 17:26:09.629843 systemd-logind[2030]: Session 6 logged out. Waiting for processes to exit. Mar 17 17:26:09.632427 systemd[1]: session-6.scope: Deactivated successfully. Mar 17 17:26:09.634612 systemd-logind[2030]: Removed session 6. Mar 17 17:26:09.652212 systemd[1]: Started sshd@6-172.31.31.32:22-139.178.68.195:58794.service - OpenSSH per-connection server daemon (139.178.68.195:58794). Mar 17 17:26:09.833386 sshd[2397]: Accepted publickey for core from 139.178.68.195 port 58794 ssh2: RSA SHA256:d/UruLZo/CsfcUUCH/x/bM9fcZFMuRhcbrxztEEs5OE Mar 17 17:26:09.835806 sshd-session[2397]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:26:09.844217 systemd-logind[2030]: New session 7 of user core. Mar 17 17:26:09.851219 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 17 17:26:09.955849 sudo[2401]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 17 17:26:09.956514 sudo[2401]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:26:10.972866 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:26:10.987220 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:26:11.030344 systemd[1]: Reloading requested from client PID 2439 ('systemctl') (unit session-7.scope)... Mar 17 17:26:11.030382 systemd[1]: Reloading... Mar 17 17:26:11.256324 zram_generator::config[2482]: No configuration found. Mar 17 17:26:11.504510 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:26:11.657629 systemd[1]: Reloading finished in 626 ms. Mar 17 17:26:11.748243 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 17 17:26:11.748457 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 17 17:26:11.749141 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:26:11.755368 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:26:12.051284 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:26:12.074504 (kubelet)[2554]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 17 17:26:12.156366 kubelet[2554]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:26:12.156366 kubelet[2554]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 17:26:12.156366 kubelet[2554]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:26:12.159511 kubelet[2554]: I0317 17:26:12.159414 2554 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 17:26:12.959534 kubelet[2554]: I0317 17:26:12.959455 2554 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Mar 17 17:26:12.959534 kubelet[2554]: I0317 17:26:12.959513 2554 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 17:26:12.959972 kubelet[2554]: I0317 17:26:12.959911 2554 server.go:927] "Client rotation is on, will bootstrap in background" Mar 17 17:26:12.997792 kubelet[2554]: I0317 17:26:12.996883 2554 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 17:26:13.016766 kubelet[2554]: I0317 17:26:13.016698 2554 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 17:26:13.019805 kubelet[2554]: I0317 17:26:13.019716 2554 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 17:26:13.020139 kubelet[2554]: I0317 17:26:13.019794 2554 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172.31.31.32","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Mar 17 17:26:13.020322 kubelet[2554]: I0317 17:26:13.020163 2554 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 17:26:13.020322 kubelet[2554]: I0317 17:26:13.020186 2554 container_manager_linux.go:301] "Creating device plugin manager" Mar 17 17:26:13.020521 kubelet[2554]: I0317 17:26:13.020426 2554 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:26:13.022551 kubelet[2554]: I0317 17:26:13.022492 2554 kubelet.go:400] "Attempting to sync node with API server" Mar 17 17:26:13.022551 kubelet[2554]: I0317 17:26:13.022539 2554 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 17:26:13.022706 kubelet[2554]: I0317 17:26:13.022652 2554 kubelet.go:312] "Adding apiserver pod source" Mar 17 17:26:13.022706 kubelet[2554]: I0317 17:26:13.022699 2554 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 17:26:13.023831 kubelet[2554]: E0317 17:26:13.023350 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:13.023831 kubelet[2554]: E0317 17:26:13.023489 2554 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:13.024535 kubelet[2554]: I0317 17:26:13.024507 2554 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Mar 17 17:26:13.025277 kubelet[2554]: I0317 17:26:13.025252 2554 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 17:26:13.025438 kubelet[2554]: W0317 17:26:13.025418 2554 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 17 17:26:13.027298 kubelet[2554]: I0317 17:26:13.026679 2554 server.go:1264] "Started kubelet" Mar 17 17:26:13.031060 kubelet[2554]: I0317 17:26:13.030945 2554 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 17:26:13.032932 kubelet[2554]: I0317 17:26:13.032130 2554 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 17:26:13.032932 kubelet[2554]: I0317 17:26:13.032227 2554 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 17:26:13.032932 kubelet[2554]: I0317 17:26:13.032389 2554 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 17:26:13.033995 kubelet[2554]: I0317 17:26:13.033938 2554 server.go:455] "Adding debug handlers to kubelet server" Mar 17 17:26:13.042103 kubelet[2554]: I0317 17:26:13.041014 2554 volume_manager.go:291] "Starting Kubelet Volume Manager" Mar 17 17:26:13.042947 kubelet[2554]: I0317 17:26:13.042627 2554 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 17 17:26:13.045821 kubelet[2554]: I0317 17:26:13.044559 2554 reconciler.go:26] "Reconciler: start to sync state" Mar 17 17:26:13.046835 kubelet[2554]: E0317 17:26:13.046801 2554 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 17:26:13.048721 kubelet[2554]: I0317 17:26:13.048686 2554 factory.go:221] Registration of the systemd container factory successfully Mar 17 17:26:13.049051 kubelet[2554]: I0317 17:26:13.049020 2554 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 17:26:13.056307 kubelet[2554]: I0317 17:26:13.056216 2554 factory.go:221] Registration of the containerd container factory successfully Mar 17 17:26:13.100282 kubelet[2554]: E0317 17:26:13.100051 2554 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.31.32.182da71cce6f694e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.31.32,UID:172.31.31.32,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172.31.31.32,},FirstTimestamp:2025-03-17 17:26:13.026580814 +0000 UTC m=+0.945002898,LastTimestamp:2025-03-17 17:26:13.026580814 +0000 UTC m=+0.945002898,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.31.32,}" Mar 17 17:26:13.101158 kubelet[2554]: W0317 17:26:13.101096 2554 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "172.31.31.32" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 17 17:26:13.101335 kubelet[2554]: E0317 17:26:13.101313 2554 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes "172.31.31.32" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 17 17:26:13.101557 kubelet[2554]: W0317 17:26:13.101526 2554 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 17 17:26:13.101681 kubelet[2554]: E0317 17:26:13.101658 2554 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 17 17:26:13.101912 kubelet[2554]: W0317 17:26:13.101882 2554 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 17 17:26:13.102057 kubelet[2554]: E0317 17:26:13.102036 2554 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 17 17:26:13.106801 kubelet[2554]: E0317 17:26:13.105333 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"172.31.31.32\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Mar 17 17:26:13.115753 kubelet[2554]: I0317 17:26:13.115691 2554 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 17:26:13.115950 kubelet[2554]: I0317 17:26:13.115924 2554 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 17:26:13.116112 kubelet[2554]: I0317 17:26:13.116090 2554 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:26:13.122273 kubelet[2554]: I0317 17:26:13.122236 2554 policy_none.go:49] "None policy: Start" Mar 17 17:26:13.125975 kubelet[2554]: E0317 17:26:13.125826 2554 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.31.32.182da71ccfa399ca default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.31.32,UID:172.31.31.32,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:172.31.31.32,},FirstTimestamp:2025-03-17 17:26:13.046778314 +0000 UTC m=+0.965200434,LastTimestamp:2025-03-17 17:26:13.046778314 +0000 UTC m=+0.965200434,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.31.32,}" Mar 17 17:26:13.127122 kubelet[2554]: I0317 17:26:13.127066 2554 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 17:26:13.127122 kubelet[2554]: I0317 17:26:13.127121 2554 state_mem.go:35] "Initializing new in-memory state store" Mar 17 17:26:13.142786 kubelet[2554]: I0317 17:26:13.141784 2554 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 17:26:13.142786 kubelet[2554]: I0317 17:26:13.142092 2554 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 17:26:13.144553 kubelet[2554]: I0317 17:26:13.144462 2554 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 17:26:13.154497 kubelet[2554]: I0317 17:26:13.154428 2554 kubelet_node_status.go:73] "Attempting to register node" node="172.31.31.32" Mar 17 17:26:13.158808 kubelet[2554]: E0317 17:26:13.158762 2554 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172.31.31.32\" not found" Mar 17 17:26:13.170063 kubelet[2554]: I0317 17:26:13.170011 2554 kubelet_node_status.go:76] "Successfully registered node" node="172.31.31.32" Mar 17 17:26:13.172088 kubelet[2554]: I0317 17:26:13.172038 2554 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 17:26:13.174378 kubelet[2554]: I0317 17:26:13.174338 2554 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 17:26:13.174622 kubelet[2554]: I0317 17:26:13.174600 2554 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 17:26:13.174864 kubelet[2554]: I0317 17:26:13.174720 2554 kubelet.go:2337] "Starting kubelet main sync loop" Mar 17 17:26:13.175047 kubelet[2554]: E0317 17:26:13.175022 2554 kubelet.go:2361] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Mar 17 17:26:13.224547 kubelet[2554]: E0317 17:26:13.224382 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.325263 kubelet[2554]: E0317 17:26:13.325160 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.426191 kubelet[2554]: E0317 17:26:13.426134 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.527330 kubelet[2554]: E0317 17:26:13.527193 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.627983 kubelet[2554]: E0317 17:26:13.627912 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.728873 kubelet[2554]: E0317 17:26:13.728812 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.829847 kubelet[2554]: E0317 17:26:13.829677 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.930624 kubelet[2554]: E0317 17:26:13.930556 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:13.963350 kubelet[2554]: I0317 17:26:13.963273 2554 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 17 17:26:13.963567 kubelet[2554]: W0317 17:26:13.963518 2554 reflector.go:470] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Mar 17 17:26:14.023821 kubelet[2554]: E0317 17:26:14.023712 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:14.031132 kubelet[2554]: E0317 17:26:14.031069 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:14.096412 sudo[2401]: pam_unix(sudo:session): session closed for user root Mar 17 17:26:14.119625 sshd[2400]: Connection closed by 139.178.68.195 port 58794 Mar 17 17:26:14.120511 sshd-session[2397]: pam_unix(sshd:session): session closed for user core Mar 17 17:26:14.128115 systemd[1]: sshd@6-172.31.31.32:22-139.178.68.195:58794.service: Deactivated successfully. Mar 17 17:26:14.133036 kubelet[2554]: E0317 17:26:14.131153 2554 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"172.31.31.32\" not found" Mar 17 17:26:14.134633 systemd-logind[2030]: Session 7 logged out. Waiting for processes to exit. Mar 17 17:26:14.134641 systemd[1]: session-7.scope: Deactivated successfully. Mar 17 17:26:14.139354 systemd-logind[2030]: Removed session 7. Mar 17 17:26:14.233424 kubelet[2554]: I0317 17:26:14.233133 2554 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Mar 17 17:26:14.234122 containerd[2072]: time="2025-03-17T17:26:14.233806807Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 17 17:26:14.234699 kubelet[2554]: I0317 17:26:14.234162 2554 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Mar 17 17:26:15.024598 kubelet[2554]: I0317 17:26:15.024464 2554 apiserver.go:52] "Watching apiserver" Mar 17 17:26:15.024598 kubelet[2554]: E0317 17:26:15.024476 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:15.032300 kubelet[2554]: I0317 17:26:15.032107 2554 topology_manager.go:215] "Topology Admit Handler" podUID="cddbf24f-1f90-4612-b331-56b3a73cdb3f" podNamespace="calico-system" podName="calico-node-9qvxd" Mar 17 17:26:15.032300 kubelet[2554]: I0317 17:26:15.032262 2554 topology_manager.go:215] "Topology Admit Handler" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" podNamespace="calico-system" podName="csi-node-driver-g9rjb" Mar 17 17:26:15.032300 kubelet[2554]: I0317 17:26:15.032377 2554 topology_manager.go:215] "Topology Admit Handler" podUID="615fd325-8715-4432-b6ef-09caa9a0320d" podNamespace="kube-system" podName="kube-proxy-wbwkp" Mar 17 17:26:15.032300 kubelet[2554]: E0317 17:26:15.032615 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:15.047528 kubelet[2554]: I0317 17:26:15.047491 2554 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 17 17:26:15.056162 kubelet[2554]: I0317 17:26:15.056117 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/615fd325-8715-4432-b6ef-09caa9a0320d-kube-proxy\") pod \"kube-proxy-wbwkp\" (UID: \"615fd325-8715-4432-b6ef-09caa9a0320d\") " pod="kube-system/kube-proxy-wbwkp" Mar 17 17:26:15.056475 kubelet[2554]: I0317 17:26:15.056447 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/615fd325-8715-4432-b6ef-09caa9a0320d-lib-modules\") pod \"kube-proxy-wbwkp\" (UID: \"615fd325-8715-4432-b6ef-09caa9a0320d\") " pod="kube-system/kube-proxy-wbwkp" Mar 17 17:26:15.056612 kubelet[2554]: I0317 17:26:15.056587 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqchf\" (UniqueName: \"kubernetes.io/projected/615fd325-8715-4432-b6ef-09caa9a0320d-kube-api-access-jqchf\") pod \"kube-proxy-wbwkp\" (UID: \"615fd325-8715-4432-b6ef-09caa9a0320d\") " pod="kube-system/kube-proxy-wbwkp" Mar 17 17:26:15.056764 kubelet[2554]: I0317 17:26:15.056715 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-lib-modules\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.056935 kubelet[2554]: I0317 17:26:15.056895 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/d2efaf52-bf0a-463d-906a-d3ef9134a72b-varrun\") pod \"csi-node-driver-g9rjb\" (UID: \"d2efaf52-bf0a-463d-906a-d3ef9134a72b\") " pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:15.057022 kubelet[2554]: I0317 17:26:15.056974 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d2efaf52-bf0a-463d-906a-d3ef9134a72b-socket-dir\") pod \"csi-node-driver-g9rjb\" (UID: \"d2efaf52-bf0a-463d-906a-d3ef9134a72b\") " pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:15.057076 kubelet[2554]: I0317 17:26:15.057018 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-flexvol-driver-host\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057128 kubelet[2554]: I0317 17:26:15.057080 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6rpj\" (UniqueName: \"kubernetes.io/projected/cddbf24f-1f90-4612-b331-56b3a73cdb3f-kube-api-access-s6rpj\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057203 kubelet[2554]: I0317 17:26:15.057188 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-policysync\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057310 kubelet[2554]: I0317 17:26:15.057229 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-var-run-calico\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057310 kubelet[2554]: I0317 17:26:15.057264 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-cni-net-dir\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057406 kubelet[2554]: I0317 17:26:15.057327 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d2efaf52-bf0a-463d-906a-d3ef9134a72b-registration-dir\") pod \"csi-node-driver-g9rjb\" (UID: \"d2efaf52-bf0a-463d-906a-d3ef9134a72b\") " pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:15.057406 kubelet[2554]: I0317 17:26:15.057375 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drw6n\" (UniqueName: \"kubernetes.io/projected/d2efaf52-bf0a-463d-906a-d3ef9134a72b-kube-api-access-drw6n\") pod \"csi-node-driver-g9rjb\" (UID: \"d2efaf52-bf0a-463d-906a-d3ef9134a72b\") " pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:15.057498 kubelet[2554]: I0317 17:26:15.057410 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-xtables-lock\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057498 kubelet[2554]: I0317 17:26:15.057444 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/cddbf24f-1f90-4612-b331-56b3a73cdb3f-node-certs\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057498 kubelet[2554]: I0317 17:26:15.057483 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-cni-log-dir\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057641 kubelet[2554]: I0317 17:26:15.057519 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2efaf52-bf0a-463d-906a-d3ef9134a72b-kubelet-dir\") pod \"csi-node-driver-g9rjb\" (UID: \"d2efaf52-bf0a-463d-906a-d3ef9134a72b\") " pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:15.057641 kubelet[2554]: I0317 17:26:15.057580 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/615fd325-8715-4432-b6ef-09caa9a0320d-xtables-lock\") pod \"kube-proxy-wbwkp\" (UID: \"615fd325-8715-4432-b6ef-09caa9a0320d\") " pod="kube-system/kube-proxy-wbwkp" Mar 17 17:26:15.057641 kubelet[2554]: I0317 17:26:15.057618 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cddbf24f-1f90-4612-b331-56b3a73cdb3f-tigera-ca-bundle\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057828 kubelet[2554]: I0317 17:26:15.057663 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-var-lib-calico\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.057828 kubelet[2554]: I0317 17:26:15.057697 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/cddbf24f-1f90-4612-b331-56b3a73cdb3f-cni-bin-dir\") pod \"calico-node-9qvxd\" (UID: \"cddbf24f-1f90-4612-b331-56b3a73cdb3f\") " pod="calico-system/calico-node-9qvxd" Mar 17 17:26:15.160611 kubelet[2554]: E0317 17:26:15.160541 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.160905 kubelet[2554]: W0317 17:26:15.160801 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.161373 kubelet[2554]: E0317 17:26:15.161342 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.161764 kubelet[2554]: E0317 17:26:15.161639 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.161968 kubelet[2554]: W0317 17:26:15.161935 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.162143 kubelet[2554]: E0317 17:26:15.162099 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.162789 kubelet[2554]: E0317 17:26:15.162661 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.162789 kubelet[2554]: W0317 17:26:15.162693 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.162789 kubelet[2554]: E0317 17:26:15.162762 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.163329 kubelet[2554]: E0317 17:26:15.163269 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.163329 kubelet[2554]: W0317 17:26:15.163327 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.165770 kubelet[2554]: E0317 17:26:15.163966 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.165770 kubelet[2554]: E0317 17:26:15.164353 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.165770 kubelet[2554]: W0317 17:26:15.164386 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.165770 kubelet[2554]: E0317 17:26:15.165383 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.165770 kubelet[2554]: W0317 17:26:15.165415 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.166569 kubelet[2554]: E0317 17:26:15.166509 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.166569 kubelet[2554]: W0317 17:26:15.166554 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.167190 kubelet[2554]: E0317 17:26:15.167127 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.167298 kubelet[2554]: W0317 17:26:15.167168 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.168196 kubelet[2554]: E0317 17:26:15.168119 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.168196 kubelet[2554]: W0317 17:26:15.168187 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.168665 kubelet[2554]: E0317 17:26:15.168623 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.168665 kubelet[2554]: W0317 17:26:15.168657 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.169064 kubelet[2554]: E0317 17:26:15.169024 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.169064 kubelet[2554]: W0317 17:26:15.169056 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.170320 kubelet[2554]: E0317 17:26:15.170269 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.170320 kubelet[2554]: W0317 17:26:15.170308 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.170675 kubelet[2554]: E0317 17:26:15.170635 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.170675 kubelet[2554]: W0317 17:26:15.170667 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.170808 kubelet[2554]: E0317 17:26:15.170698 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.171210 kubelet[2554]: E0317 17:26:15.171160 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.171210 kubelet[2554]: W0317 17:26:15.171200 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.171378 kubelet[2554]: E0317 17:26:15.171230 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.171637 kubelet[2554]: E0317 17:26:15.171576 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.172135 kubelet[2554]: E0317 17:26:15.172068 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.172135 kubelet[2554]: W0317 17:26:15.172125 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.172289 kubelet[2554]: E0317 17:26:15.172160 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.172589 kubelet[2554]: E0317 17:26:15.172528 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.172589 kubelet[2554]: W0317 17:26:15.172580 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.172794 kubelet[2554]: E0317 17:26:15.172619 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.174790 kubelet[2554]: E0317 17:26:15.173124 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.174790 kubelet[2554]: W0317 17:26:15.173168 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.174790 kubelet[2554]: E0317 17:26:15.173204 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.179934 kubelet[2554]: E0317 17:26:15.177936 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188272 kubelet[2554]: E0317 17:26:15.180027 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188438 kubelet[2554]: E0317 17:26:15.180054 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188546 kubelet[2554]: E0317 17:26:15.180064 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188671 kubelet[2554]: E0317 17:26:15.180073 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188859 kubelet[2554]: E0317 17:26:15.180082 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.188969 kubelet[2554]: E0317 17:26:15.180091 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.189135 kubelet[2554]: E0317 17:26:15.188206 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.189259 kubelet[2554]: W0317 17:26:15.189226 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.189392 kubelet[2554]: E0317 17:26:15.189360 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.190042 kubelet[2554]: E0317 17:26:15.190007 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.190206 kubelet[2554]: W0317 17:26:15.190178 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.190325 kubelet[2554]: E0317 17:26:15.190301 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.190848 kubelet[2554]: E0317 17:26:15.190823 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.190982 kubelet[2554]: W0317 17:26:15.190958 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.191089 kubelet[2554]: E0317 17:26:15.191067 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.191539 kubelet[2554]: E0317 17:26:15.191514 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.191705 kubelet[2554]: W0317 17:26:15.191679 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.191879 kubelet[2554]: E0317 17:26:15.191856 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.192340 kubelet[2554]: E0317 17:26:15.192310 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.192495 kubelet[2554]: W0317 17:26:15.192466 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.192605 kubelet[2554]: E0317 17:26:15.192582 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.193118 kubelet[2554]: E0317 17:26:15.193088 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.193260 kubelet[2554]: W0317 17:26:15.193235 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.193401 kubelet[2554]: E0317 17:26:15.193376 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.199183 kubelet[2554]: E0317 17:26:15.198988 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.199183 kubelet[2554]: W0317 17:26:15.199032 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.199183 kubelet[2554]: E0317 17:26:15.199069 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.202430 kubelet[2554]: E0317 17:26:15.202394 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.202633 kubelet[2554]: W0317 17:26:15.202605 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.202963 kubelet[2554]: E0317 17:26:15.202723 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.203291 kubelet[2554]: E0317 17:26:15.203267 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.203411 kubelet[2554]: W0317 17:26:15.203386 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.203527 kubelet[2554]: E0317 17:26:15.203502 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.204197 kubelet[2554]: E0317 17:26:15.203963 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.204197 kubelet[2554]: W0317 17:26:15.203991 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.204197 kubelet[2554]: E0317 17:26:15.204017 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.212812 kubelet[2554]: E0317 17:26:15.212137 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.215624 kubelet[2554]: W0317 17:26:15.212973 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.215624 kubelet[2554]: E0317 17:26:15.213030 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.225770 kubelet[2554]: E0317 17:26:15.223315 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.226679 kubelet[2554]: W0317 17:26:15.226396 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.226679 kubelet[2554]: E0317 17:26:15.226450 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.227257 kubelet[2554]: E0317 17:26:15.227230 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.227377 kubelet[2554]: W0317 17:26:15.227351 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.227511 kubelet[2554]: E0317 17:26:15.227488 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.229141 kubelet[2554]: E0317 17:26:15.228931 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.229141 kubelet[2554]: W0317 17:26:15.228963 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.229141 kubelet[2554]: E0317 17:26:15.229010 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.230266 kubelet[2554]: E0317 17:26:15.230148 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.230266 kubelet[2554]: W0317 17:26:15.230177 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.230266 kubelet[2554]: E0317 17:26:15.230221 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.232259 kubelet[2554]: E0317 17:26:15.232200 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.232441 kubelet[2554]: W0317 17:26:15.232266 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.232441 kubelet[2554]: E0317 17:26:15.232318 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.233041 kubelet[2554]: E0317 17:26:15.232984 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.233182 kubelet[2554]: W0317 17:26:15.233063 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.233398 kubelet[2554]: E0317 17:26:15.233232 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.233682 kubelet[2554]: E0317 17:26:15.233629 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.234569 kubelet[2554]: W0317 17:26:15.233683 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.234569 kubelet[2554]: E0317 17:26:15.233785 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.234569 kubelet[2554]: E0317 17:26:15.234228 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.234569 kubelet[2554]: W0317 17:26:15.234252 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.234569 kubelet[2554]: E0317 17:26:15.234333 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.235060 kubelet[2554]: E0317 17:26:15.234803 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.235060 kubelet[2554]: W0317 17:26:15.234828 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.235060 kubelet[2554]: E0317 17:26:15.234911 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.235298 kubelet[2554]: E0317 17:26:15.235268 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.235389 kubelet[2554]: W0317 17:26:15.235297 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.235433 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.235686 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.236929 kubelet[2554]: W0317 17:26:15.235708 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.235786 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.236161 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.236929 kubelet[2554]: W0317 17:26:15.236181 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.236203 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.236638 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.236929 kubelet[2554]: W0317 17:26:15.236688 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.236929 kubelet[2554]: E0317 17:26:15.236716 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.237626 kubelet[2554]: E0317 17:26:15.237566 2554 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:26:15.237626 kubelet[2554]: W0317 17:26:15.237625 2554 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:26:15.237872 kubelet[2554]: E0317 17:26:15.237659 2554 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:26:15.340717 containerd[2072]: time="2025-03-17T17:26:15.340556279Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9qvxd,Uid:cddbf24f-1f90-4612-b331-56b3a73cdb3f,Namespace:calico-system,Attempt:0,}" Mar 17 17:26:15.345553 containerd[2072]: time="2025-03-17T17:26:15.345374460Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wbwkp,Uid:615fd325-8715-4432-b6ef-09caa9a0320d,Namespace:kube-system,Attempt:0,}" Mar 17 17:26:15.951548 containerd[2072]: time="2025-03-17T17:26:15.951469075Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:26:15.956021 containerd[2072]: time="2025-03-17T17:26:15.955945308Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Mar 17 17:26:15.959698 containerd[2072]: time="2025-03-17T17:26:15.958957121Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:26:15.961561 containerd[2072]: time="2025-03-17T17:26:15.961513055Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:26:15.963400 containerd[2072]: time="2025-03-17T17:26:15.963330625Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 17 17:26:15.967464 containerd[2072]: time="2025-03-17T17:26:15.967399854Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:26:15.969660 containerd[2072]: time="2025-03-17T17:26:15.969334378Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 623.23434ms" Mar 17 17:26:15.974060 containerd[2072]: time="2025-03-17T17:26:15.973826383Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 633.141169ms" Mar 17 17:26:16.024945 kubelet[2554]: E0317 17:26:16.024836 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:16.176206 kubelet[2554]: E0317 17:26:16.176147 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:16.189785 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2251102423.mount: Deactivated successfully. Mar 17 17:26:16.226182 containerd[2072]: time="2025-03-17T17:26:16.225314141Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:26:16.229594 containerd[2072]: time="2025-03-17T17:26:16.229259112Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:26:16.230575 containerd[2072]: time="2025-03-17T17:26:16.229945960Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:26:16.230575 containerd[2072]: time="2025-03-17T17:26:16.230334434Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:26:16.230575 containerd[2072]: time="2025-03-17T17:26:16.230470745Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:16.231196 containerd[2072]: time="2025-03-17T17:26:16.230962739Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:16.231286 containerd[2072]: time="2025-03-17T17:26:16.231134182Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:16.231286 containerd[2072]: time="2025-03-17T17:26:16.231084418Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:16.440186 containerd[2072]: time="2025-03-17T17:26:16.439952601Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wbwkp,Uid:615fd325-8715-4432-b6ef-09caa9a0320d,Namespace:kube-system,Attempt:0,} returns sandbox id \"3ba4e41f1de88e520f24947879c068dd45084e49f465738d459499d1aa7afabe\"" Mar 17 17:26:16.444775 containerd[2072]: time="2025-03-17T17:26:16.444230944Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-9qvxd,Uid:cddbf24f-1f90-4612-b331-56b3a73cdb3f,Namespace:calico-system,Attempt:0,} returns sandbox id \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\"" Mar 17 17:26:16.448204 containerd[2072]: time="2025-03-17T17:26:16.448027117Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\"" Mar 17 17:26:17.025053 kubelet[2554]: E0317 17:26:17.025009 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:17.879779 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount640153345.mount: Deactivated successfully. Mar 17 17:26:18.012434 containerd[2072]: time="2025-03-17T17:26:18.012357370Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:18.015053 containerd[2072]: time="2025-03-17T17:26:18.014945339Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2: active requests=0, bytes read=6490047" Mar 17 17:26:18.017612 containerd[2072]: time="2025-03-17T17:26:18.017528439Z" level=info msg="ImageCreate event name:\"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:18.022500 containerd[2072]: time="2025-03-17T17:26:18.022420204Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:18.025458 containerd[2072]: time="2025-03-17T17:26:18.025326194Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" with image id \"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\", size \"6489869\" in 1.576760822s" Mar 17 17:26:18.025458 containerd[2072]: time="2025-03-17T17:26:18.025399058Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" returns image reference \"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\"" Mar 17 17:26:18.025665 kubelet[2554]: E0317 17:26:18.025509 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:18.031573 containerd[2072]: time="2025-03-17T17:26:18.031493593Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\"" Mar 17 17:26:18.034626 containerd[2072]: time="2025-03-17T17:26:18.034562845Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 17 17:26:18.068163 containerd[2072]: time="2025-03-17T17:26:18.068088786Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e\"" Mar 17 17:26:18.069327 containerd[2072]: time="2025-03-17T17:26:18.069273074Z" level=info msg="StartContainer for \"1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e\"" Mar 17 17:26:18.176952 kubelet[2554]: E0317 17:26:18.176680 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:18.179218 containerd[2072]: time="2025-03-17T17:26:18.179153502Z" level=info msg="StartContainer for \"1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e\" returns successfully" Mar 17 17:26:18.277290 containerd[2072]: time="2025-03-17T17:26:18.277080087Z" level=info msg="shim disconnected" id=1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e namespace=k8s.io Mar 17 17:26:18.277290 containerd[2072]: time="2025-03-17T17:26:18.277232711Z" level=warning msg="cleaning up after shim disconnected" id=1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e namespace=k8s.io Mar 17 17:26:18.277936 containerd[2072]: time="2025-03-17T17:26:18.277254420Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 17 17:26:18.832400 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1f1eace7ab434591401a259ba62732747d930c8938c6aa1138e70178d0f94c6e-rootfs.mount: Deactivated successfully. Mar 17 17:26:19.026321 kubelet[2554]: E0317 17:26:19.026214 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:19.317935 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount864899342.mount: Deactivated successfully. Mar 17 17:26:19.834225 containerd[2072]: time="2025-03-17T17:26:19.834168638Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:19.837086 containerd[2072]: time="2025-03-17T17:26:19.837029075Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.11: active requests=0, bytes read=25771848" Mar 17 17:26:19.838561 containerd[2072]: time="2025-03-17T17:26:19.838515551Z" level=info msg="ImageCreate event name:\"sha256:fe83790bf8a35411788b67fe5f0ce35309056c40530484d516af2ca01375220c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:19.841923 containerd[2072]: time="2025-03-17T17:26:19.841861457Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea4da798040a18ed3f302e8d5f67307c7275a2a53bcf3d51bcec223acda84a55\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:19.843587 containerd[2072]: time="2025-03-17T17:26:19.843531190Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.11\" with image id \"sha256:fe83790bf8a35411788b67fe5f0ce35309056c40530484d516af2ca01375220c\", repo tag \"registry.k8s.io/kube-proxy:v1.30.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea4da798040a18ed3f302e8d5f67307c7275a2a53bcf3d51bcec223acda84a55\", size \"25770867\" in 1.811957466s" Mar 17 17:26:19.843718 containerd[2072]: time="2025-03-17T17:26:19.843583808Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\" returns image reference \"sha256:fe83790bf8a35411788b67fe5f0ce35309056c40530484d516af2ca01375220c\"" Mar 17 17:26:19.847109 containerd[2072]: time="2025-03-17T17:26:19.847064299Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\"" Mar 17 17:26:19.848493 containerd[2072]: time="2025-03-17T17:26:19.848441270Z" level=info msg="CreateContainer within sandbox \"3ba4e41f1de88e520f24947879c068dd45084e49f465738d459499d1aa7afabe\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 17 17:26:19.873020 containerd[2072]: time="2025-03-17T17:26:19.872948778Z" level=info msg="CreateContainer within sandbox \"3ba4e41f1de88e520f24947879c068dd45084e49f465738d459499d1aa7afabe\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"e8291a2741cfad7779c53a3a2dbdbebb025889e500ebd6851d31bef6d0fcd493\"" Mar 17 17:26:19.875828 containerd[2072]: time="2025-03-17T17:26:19.874130655Z" level=info msg="StartContainer for \"e8291a2741cfad7779c53a3a2dbdbebb025889e500ebd6851d31bef6d0fcd493\"" Mar 17 17:26:19.989758 containerd[2072]: time="2025-03-17T17:26:19.989670932Z" level=info msg="StartContainer for \"e8291a2741cfad7779c53a3a2dbdbebb025889e500ebd6851d31bef6d0fcd493\" returns successfully" Mar 17 17:26:20.026929 kubelet[2554]: E0317 17:26:20.026844 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:20.175701 kubelet[2554]: E0317 17:26:20.175605 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:21.027558 kubelet[2554]: E0317 17:26:21.027486 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:22.028670 kubelet[2554]: E0317 17:26:22.028510 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:22.176689 kubelet[2554]: E0317 17:26:22.176321 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:23.028795 kubelet[2554]: E0317 17:26:23.028646 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:23.384917 containerd[2072]: time="2025-03-17T17:26:23.384841584Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:23.386833 containerd[2072]: time="2025-03-17T17:26:23.386752576Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.2: active requests=0, bytes read=91227396" Mar 17 17:26:23.389106 containerd[2072]: time="2025-03-17T17:26:23.389034998Z" level=info msg="ImageCreate event name:\"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:23.393917 containerd[2072]: time="2025-03-17T17:26:23.393827165Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:23.396205 containerd[2072]: time="2025-03-17T17:26:23.395294535Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.2\" with image id \"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\", size \"92597153\" in 3.54801053s" Mar 17 17:26:23.396205 containerd[2072]: time="2025-03-17T17:26:23.395347296Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\" returns image reference \"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\"" Mar 17 17:26:23.399102 containerd[2072]: time="2025-03-17T17:26:23.398925202Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 17 17:26:23.428025 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1757785681.mount: Deactivated successfully. Mar 17 17:26:23.433232 containerd[2072]: time="2025-03-17T17:26:23.433162028Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf\"" Mar 17 17:26:23.434008 containerd[2072]: time="2025-03-17T17:26:23.433715431Z" level=info msg="StartContainer for \"14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf\"" Mar 17 17:26:23.544287 containerd[2072]: time="2025-03-17T17:26:23.544186408Z" level=info msg="StartContainer for \"14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf\" returns successfully" Mar 17 17:26:24.030774 kubelet[2554]: E0317 17:26:24.029687 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:24.176074 kubelet[2554]: E0317 17:26:24.175994 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:24.303545 kubelet[2554]: I0317 17:26:24.303226 2554 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wbwkp" podStartSLOduration=7.905169045 podStartE2EDuration="11.303203324s" podCreationTimestamp="2025-03-17 17:26:13 +0000 UTC" firstStartedPulling="2025-03-17 17:26:16.448149492 +0000 UTC m=+4.366571564" lastFinishedPulling="2025-03-17 17:26:19.846183735 +0000 UTC m=+7.764605843" observedRunningTime="2025-03-17 17:26:20.301382267 +0000 UTC m=+8.219804387" watchObservedRunningTime="2025-03-17 17:26:24.303203324 +0000 UTC m=+12.221625408" Mar 17 17:26:24.412373 containerd[2072]: time="2025-03-17T17:26:24.412304850Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 17:26:24.455851 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf-rootfs.mount: Deactivated successfully. Mar 17 17:26:24.466472 kubelet[2554]: I0317 17:26:24.466404 2554 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Mar 17 17:26:25.029903 kubelet[2554]: E0317 17:26:25.029825 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:26.030633 kubelet[2554]: E0317 17:26:26.030552 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:26.081261 containerd[2072]: time="2025-03-17T17:26:26.081159694Z" level=info msg="shim disconnected" id=14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf namespace=k8s.io Mar 17 17:26:26.081261 containerd[2072]: time="2025-03-17T17:26:26.081245787Z" level=warning msg="cleaning up after shim disconnected" id=14d5dae6b6c1752f2b8926b5d13d81b4e52a29304751bdc3b8ff9cd30ae45baf namespace=k8s.io Mar 17 17:26:26.081261 containerd[2072]: time="2025-03-17T17:26:26.081267496Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 17 17:26:26.182038 containerd[2072]: time="2025-03-17T17:26:26.181520725Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:0,}" Mar 17 17:26:26.280810 containerd[2072]: time="2025-03-17T17:26:26.279840461Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\"" Mar 17 17:26:26.313355 containerd[2072]: time="2025-03-17T17:26:26.313294953Z" level=error msg="Failed to destroy network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:26.317252 containerd[2072]: time="2025-03-17T17:26:26.316183924Z" level=error msg="encountered an error cleaning up failed sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:26.317252 containerd[2072]: time="2025-03-17T17:26:26.316284746Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:26.317569 kubelet[2554]: E0317 17:26:26.316742 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:26.317569 kubelet[2554]: E0317 17:26:26.316833 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:26.317569 kubelet[2554]: E0317 17:26:26.316868 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:26.317837 kubelet[2554]: E0317 17:26:26.316930 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:26.318138 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820-shm.mount: Deactivated successfully. Mar 17 17:26:27.031279 kubelet[2554]: E0317 17:26:27.031213 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:27.281687 kubelet[2554]: I0317 17:26:27.281499 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820" Mar 17 17:26:27.283040 containerd[2072]: time="2025-03-17T17:26:27.282953904Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:27.283664 containerd[2072]: time="2025-03-17T17:26:27.283365250Z" level=info msg="Ensure that sandbox be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820 in task-service has been cleanup successfully" Mar 17 17:26:27.283975 containerd[2072]: time="2025-03-17T17:26:27.283759925Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:27.283975 containerd[2072]: time="2025-03-17T17:26:27.283797646Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:27.286923 systemd[1]: run-netns-cni\x2db412f052\x2dfade\x2d3e69\x2d4316\x2d806df7ba0a47.mount: Deactivated successfully. Mar 17 17:26:27.287554 containerd[2072]: time="2025-03-17T17:26:27.287357094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:1,}" Mar 17 17:26:27.398675 containerd[2072]: time="2025-03-17T17:26:27.398617780Z" level=error msg="Failed to destroy network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:27.399615 containerd[2072]: time="2025-03-17T17:26:27.399449899Z" level=error msg="encountered an error cleaning up failed sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:27.399615 containerd[2072]: time="2025-03-17T17:26:27.399547830Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:27.400289 kubelet[2554]: E0317 17:26:27.400131 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:27.400289 kubelet[2554]: E0317 17:26:27.400234 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:27.400741 kubelet[2554]: E0317 17:26:27.400471 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:27.401234 kubelet[2554]: E0317 17:26:27.401002 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:27.403675 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e-shm.mount: Deactivated successfully. Mar 17 17:26:27.735766 kubelet[2554]: I0317 17:26:27.734916 2554 topology_manager.go:215] "Topology Admit Handler" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" podNamespace="default" podName="nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:27.750273 kubelet[2554]: I0317 17:26:27.750213 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ww94\" (UniqueName: \"kubernetes.io/projected/44cae95b-023f-401f-ab12-ec5a58221ec7-kube-api-access-9ww94\") pod \"nginx-deployment-85f456d6dd-jshxr\" (UID: \"44cae95b-023f-401f-ab12-ec5a58221ec7\") " pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:28.032695 kubelet[2554]: E0317 17:26:28.032542 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:28.059874 containerd[2072]: time="2025-03-17T17:26:28.059809376Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:0,}" Mar 17 17:26:28.236985 containerd[2072]: time="2025-03-17T17:26:28.236897081Z" level=error msg="Failed to destroy network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.238008 containerd[2072]: time="2025-03-17T17:26:28.237890940Z" level=error msg="encountered an error cleaning up failed sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.238291 containerd[2072]: time="2025-03-17T17:26:28.238137272Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:0,} failed, error" error="failed to setup network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.238861 kubelet[2554]: E0317 17:26:28.238705 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.239005 kubelet[2554]: E0317 17:26:28.238861 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:28.239005 kubelet[2554]: E0317 17:26:28.238896 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:28.239005 kubelet[2554]: E0317 17:26:28.238975 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:28.291889 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa-shm.mount: Deactivated successfully. Mar 17 17:26:28.294055 kubelet[2554]: I0317 17:26:28.292110 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e" Mar 17 17:26:28.296171 containerd[2072]: time="2025-03-17T17:26:28.296116952Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:28.298268 containerd[2072]: time="2025-03-17T17:26:28.297628940Z" level=info msg="Ensure that sandbox a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e in task-service has been cleanup successfully" Mar 17 17:26:28.300918 containerd[2072]: time="2025-03-17T17:26:28.300867236Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:28.301292 containerd[2072]: time="2025-03-17T17:26:28.301086534Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:28.303474 containerd[2072]: time="2025-03-17T17:26:28.303335625Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:28.303407 systemd[1]: run-netns-cni\x2da5195d90\x2d041e\x2df951\x2defe3\x2d53e283fd611d.mount: Deactivated successfully. Mar 17 17:26:28.305716 containerd[2072]: time="2025-03-17T17:26:28.304024237Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:28.305716 containerd[2072]: time="2025-03-17T17:26:28.304070078Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:28.306468 kubelet[2554]: I0317 17:26:28.304239 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa" Mar 17 17:26:28.307628 containerd[2072]: time="2025-03-17T17:26:28.307554599Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:28.308120 containerd[2072]: time="2025-03-17T17:26:28.308067990Z" level=info msg="Ensure that sandbox 59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa in task-service has been cleanup successfully" Mar 17 17:26:28.308868 containerd[2072]: time="2025-03-17T17:26:28.308773945Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:2,}" Mar 17 17:26:28.311195 containerd[2072]: time="2025-03-17T17:26:28.310980337Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:28.311195 containerd[2072]: time="2025-03-17T17:26:28.311025914Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:28.314163 containerd[2072]: time="2025-03-17T17:26:28.312784738Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:1,}" Mar 17 17:26:28.313995 systemd[1]: run-netns-cni\x2d886e986e\x2d556c\x2d2e85\x2d0335\x2de0b842cbaf91.mount: Deactivated successfully. Mar 17 17:26:28.581589 containerd[2072]: time="2025-03-17T17:26:28.581104960Z" level=error msg="Failed to destroy network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.581834 containerd[2072]: time="2025-03-17T17:26:28.581778555Z" level=error msg="encountered an error cleaning up failed sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.581917 containerd[2072]: time="2025-03-17T17:26:28.581881104Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:1,} failed, error" error="failed to setup network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.583255 kubelet[2554]: E0317 17:26:28.582188 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.583255 kubelet[2554]: E0317 17:26:28.582270 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:28.583255 kubelet[2554]: E0317 17:26:28.582335 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:28.583459 kubelet[2554]: E0317 17:26:28.582402 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:28.596111 containerd[2072]: time="2025-03-17T17:26:28.596039757Z" level=error msg="Failed to destroy network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.598673 containerd[2072]: time="2025-03-17T17:26:28.598544536Z" level=error msg="encountered an error cleaning up failed sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.598673 containerd[2072]: time="2025-03-17T17:26:28.598656512Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.599614 kubelet[2554]: E0317 17:26:28.598996 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:28.599614 kubelet[2554]: E0317 17:26:28.599068 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:28.599614 kubelet[2554]: E0317 17:26:28.599101 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:28.599919 kubelet[2554]: E0317 17:26:28.599171 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:29.033816 kubelet[2554]: E0317 17:26:29.033184 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:29.294536 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24-shm.mount: Deactivated successfully. Mar 17 17:26:29.313817 kubelet[2554]: I0317 17:26:29.313541 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8" Mar 17 17:26:29.315264 containerd[2072]: time="2025-03-17T17:26:29.315202398Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:29.318023 containerd[2072]: time="2025-03-17T17:26:29.315517240Z" level=info msg="Ensure that sandbox f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8 in task-service has been cleanup successfully" Mar 17 17:26:29.320946 containerd[2072]: time="2025-03-17T17:26:29.319711854Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:29.320946 containerd[2072]: time="2025-03-17T17:26:29.319783182Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:29.323372 systemd[1]: run-netns-cni\x2d002ec505\x2d09b9\x2d65d6\x2d69a0\x2d153b1ffac6b9.mount: Deactivated successfully. Mar 17 17:26:29.326161 containerd[2072]: time="2025-03-17T17:26:29.326091426Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:29.326311 containerd[2072]: time="2025-03-17T17:26:29.326266059Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:29.326311 containerd[2072]: time="2025-03-17T17:26:29.326293345Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:29.329185 containerd[2072]: time="2025-03-17T17:26:29.328918328Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:2,}" Mar 17 17:26:29.329762 kubelet[2554]: I0317 17:26:29.329632 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24" Mar 17 17:26:29.332108 containerd[2072]: time="2025-03-17T17:26:29.331366136Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:29.332108 containerd[2072]: time="2025-03-17T17:26:29.331838555Z" level=info msg="Ensure that sandbox 882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24 in task-service has been cleanup successfully" Mar 17 17:26:29.335310 containerd[2072]: time="2025-03-17T17:26:29.335255214Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:29.335545 containerd[2072]: time="2025-03-17T17:26:29.335508899Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:29.337606 systemd[1]: run-netns-cni\x2d86a335d9\x2db6b1\x2d7a5b\x2d713d\x2d44285b2754a3.mount: Deactivated successfully. Mar 17 17:26:29.338992 containerd[2072]: time="2025-03-17T17:26:29.338930632Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:29.339164 containerd[2072]: time="2025-03-17T17:26:29.339121792Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:29.339297 containerd[2072]: time="2025-03-17T17:26:29.339159597Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:29.343436 containerd[2072]: time="2025-03-17T17:26:29.343271044Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:29.344571 containerd[2072]: time="2025-03-17T17:26:29.344414600Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:29.344571 containerd[2072]: time="2025-03-17T17:26:29.344472927Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:29.347756 containerd[2072]: time="2025-03-17T17:26:29.346450570Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:3,}" Mar 17 17:26:29.585110 containerd[2072]: time="2025-03-17T17:26:29.584496796Z" level=error msg="Failed to destroy network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.587366 containerd[2072]: time="2025-03-17T17:26:29.587271020Z" level=error msg="encountered an error cleaning up failed sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.588373 containerd[2072]: time="2025-03-17T17:26:29.587393275Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.588492 kubelet[2554]: E0317 17:26:29.587723 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.588492 kubelet[2554]: E0317 17:26:29.587853 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:29.588492 kubelet[2554]: E0317 17:26:29.587895 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:29.588681 kubelet[2554]: E0317 17:26:29.587973 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:29.622882 containerd[2072]: time="2025-03-17T17:26:29.622812912Z" level=error msg="Failed to destroy network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.623550 containerd[2072]: time="2025-03-17T17:26:29.623473098Z" level=error msg="encountered an error cleaning up failed sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.623822 containerd[2072]: time="2025-03-17T17:26:29.623606136Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:2,} failed, error" error="failed to setup network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.624064 kubelet[2554]: E0317 17:26:29.623964 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:29.624064 kubelet[2554]: E0317 17:26:29.624046 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:29.624373 kubelet[2554]: E0317 17:26:29.624080 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:29.624373 kubelet[2554]: E0317 17:26:29.624157 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:30.034070 kubelet[2554]: E0317 17:26:30.033782 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:30.293434 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6-shm.mount: Deactivated successfully. Mar 17 17:26:30.293861 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4-shm.mount: Deactivated successfully. Mar 17 17:26:30.341153 kubelet[2554]: I0317 17:26:30.341108 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6" Mar 17 17:26:30.344597 containerd[2072]: time="2025-03-17T17:26:30.344229682Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:26:30.345682 containerd[2072]: time="2025-03-17T17:26:30.345630749Z" level=info msg="Ensure that sandbox 43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6 in task-service has been cleanup successfully" Mar 17 17:26:30.349496 containerd[2072]: time="2025-03-17T17:26:30.349022532Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:26:30.349496 containerd[2072]: time="2025-03-17T17:26:30.349080907Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:26:30.351622 systemd[1]: run-netns-cni\x2d4bb769d8\x2d88c1\x2d82ff\x2d24de\x2ddef935d2417a.mount: Deactivated successfully. Mar 17 17:26:30.355252 containerd[2072]: time="2025-03-17T17:26:30.354899232Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:30.355252 containerd[2072]: time="2025-03-17T17:26:30.355181643Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:30.356004 containerd[2072]: time="2025-03-17T17:26:30.355227856Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:30.356130 kubelet[2554]: I0317 17:26:30.355673 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4" Mar 17 17:26:30.356716 containerd[2072]: time="2025-03-17T17:26:30.356641385Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:26:30.358580 containerd[2072]: time="2025-03-17T17:26:30.357708323Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:30.358943 containerd[2072]: time="2025-03-17T17:26:30.358722740Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:30.358943 containerd[2072]: time="2025-03-17T17:26:30.358792617Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:30.360116 containerd[2072]: time="2025-03-17T17:26:30.359643219Z" level=info msg="Ensure that sandbox 15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4 in task-service has been cleanup successfully" Mar 17 17:26:30.363049 containerd[2072]: time="2025-03-17T17:26:30.362993359Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:26:30.363661 containerd[2072]: time="2025-03-17T17:26:30.363459290Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:26:30.364168 containerd[2072]: time="2025-03-17T17:26:30.363307338Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:30.364283 containerd[2072]: time="2025-03-17T17:26:30.364191356Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:30.364283 containerd[2072]: time="2025-03-17T17:26:30.364228657Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:30.365137 systemd[1]: run-netns-cni\x2d91662510\x2da568\x2d6416\x2d6811\x2db5c540e79ba6.mount: Deactivated successfully. Mar 17 17:26:30.367201 containerd[2072]: time="2025-03-17T17:26:30.367026078Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:30.367635 containerd[2072]: time="2025-03-17T17:26:30.367435313Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:30.367635 containerd[2072]: time="2025-03-17T17:26:30.367497382Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:30.368274 containerd[2072]: time="2025-03-17T17:26:30.368028728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:4,}" Mar 17 17:26:30.370167 containerd[2072]: time="2025-03-17T17:26:30.369273322Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:30.370167 containerd[2072]: time="2025-03-17T17:26:30.369474197Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:30.370167 containerd[2072]: time="2025-03-17T17:26:30.369500932Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:30.372490 containerd[2072]: time="2025-03-17T17:26:30.372422262Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:3,}" Mar 17 17:26:30.591542 containerd[2072]: time="2025-03-17T17:26:30.591360607Z" level=error msg="Failed to destroy network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.593698 containerd[2072]: time="2025-03-17T17:26:30.593467641Z" level=error msg="encountered an error cleaning up failed sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.593698 containerd[2072]: time="2025-03-17T17:26:30.593585278Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:3,} failed, error" error="failed to setup network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.594217 kubelet[2554]: E0317 17:26:30.593912 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.594217 kubelet[2554]: E0317 17:26:30.593992 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:30.594217 kubelet[2554]: E0317 17:26:30.594029 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:30.594585 kubelet[2554]: E0317 17:26:30.594096 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:30.638169 containerd[2072]: time="2025-03-17T17:26:30.638012542Z" level=error msg="Failed to destroy network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.640982 containerd[2072]: time="2025-03-17T17:26:30.640909464Z" level=error msg="encountered an error cleaning up failed sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.641155 containerd[2072]: time="2025-03-17T17:26:30.641019665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.641379 kubelet[2554]: E0317 17:26:30.641312 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:30.641502 kubelet[2554]: E0317 17:26:30.641402 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:30.641502 kubelet[2554]: E0317 17:26:30.641448 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:30.641610 kubelet[2554]: E0317 17:26:30.641517 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:30.755197 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Mar 17 17:26:31.035039 kubelet[2554]: E0317 17:26:31.034818 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:31.292495 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909-shm.mount: Deactivated successfully. Mar 17 17:26:31.294492 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792-shm.mount: Deactivated successfully. Mar 17 17:26:31.371069 kubelet[2554]: I0317 17:26:31.371017 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792" Mar 17 17:26:31.373972 containerd[2072]: time="2025-03-17T17:26:31.373164048Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:26:31.373972 containerd[2072]: time="2025-03-17T17:26:31.373802273Z" level=info msg="Ensure that sandbox 8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792 in task-service has been cleanup successfully" Mar 17 17:26:31.376300 containerd[2072]: time="2025-03-17T17:26:31.376227496Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:26:31.376300 containerd[2072]: time="2025-03-17T17:26:31.376285535Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:26:31.379755 containerd[2072]: time="2025-03-17T17:26:31.379149714Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:26:31.379755 containerd[2072]: time="2025-03-17T17:26:31.379335033Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:26:31.379755 containerd[2072]: time="2025-03-17T17:26:31.379362164Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:26:31.380692 systemd[1]: run-netns-cni\x2dc2ece722\x2d30b2\x2d3700\x2de23f\x2d6c20a09d6f63.mount: Deactivated successfully. Mar 17 17:26:31.385115 containerd[2072]: time="2025-03-17T17:26:31.384540165Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:31.385115 containerd[2072]: time="2025-03-17T17:26:31.384707074Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:31.385115 containerd[2072]: time="2025-03-17T17:26:31.384770726Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:31.386422 containerd[2072]: time="2025-03-17T17:26:31.386332165Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:31.386602 containerd[2072]: time="2025-03-17T17:26:31.386552842Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:31.386602 containerd[2072]: time="2025-03-17T17:26:31.386591415Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:31.388010 containerd[2072]: time="2025-03-17T17:26:31.387937682Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:31.388160 containerd[2072]: time="2025-03-17T17:26:31.388118623Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:31.388224 containerd[2072]: time="2025-03-17T17:26:31.388159427Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:31.389211 kubelet[2554]: I0317 17:26:31.388935 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909" Mar 17 17:26:31.390690 containerd[2072]: time="2025-03-17T17:26:31.390633189Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:5,}" Mar 17 17:26:31.392195 containerd[2072]: time="2025-03-17T17:26:31.391227072Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:26:31.392483 containerd[2072]: time="2025-03-17T17:26:31.392429459Z" level=info msg="Ensure that sandbox 4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909 in task-service has been cleanup successfully" Mar 17 17:26:31.395114 containerd[2072]: time="2025-03-17T17:26:31.395043419Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:26:31.395114 containerd[2072]: time="2025-03-17T17:26:31.395100654Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:26:31.397368 systemd[1]: run-netns-cni\x2d0e57a5a1\x2d861f\x2d6218\x2dd605\x2dfb4dc950d0af.mount: Deactivated successfully. Mar 17 17:26:31.401596 containerd[2072]: time="2025-03-17T17:26:31.401390776Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:26:31.401596 containerd[2072]: time="2025-03-17T17:26:31.401589229Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:26:31.401849 containerd[2072]: time="2025-03-17T17:26:31.401615411Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:26:31.403813 containerd[2072]: time="2025-03-17T17:26:31.403658433Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:31.404006 containerd[2072]: time="2025-03-17T17:26:31.403871591Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:31.404006 containerd[2072]: time="2025-03-17T17:26:31.403901564Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:31.405347 containerd[2072]: time="2025-03-17T17:26:31.405288514Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:31.405506 containerd[2072]: time="2025-03-17T17:26:31.405471578Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:31.405586 containerd[2072]: time="2025-03-17T17:26:31.405502511Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:31.408437 containerd[2072]: time="2025-03-17T17:26:31.408373455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:4,}" Mar 17 17:26:31.649056 containerd[2072]: time="2025-03-17T17:26:31.648978061Z" level=error msg="Failed to destroy network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.650083 containerd[2072]: time="2025-03-17T17:26:31.650009233Z" level=error msg="encountered an error cleaning up failed sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.650371 containerd[2072]: time="2025-03-17T17:26:31.650123620Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.650933 kubelet[2554]: E0317 17:26:31.650470 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.650933 kubelet[2554]: E0317 17:26:31.650549 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:31.650933 kubelet[2554]: E0317 17:26:31.650587 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:31.651147 kubelet[2554]: E0317 17:26:31.650663 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:31.656056 containerd[2072]: time="2025-03-17T17:26:31.655972326Z" level=error msg="Failed to destroy network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.657120 containerd[2072]: time="2025-03-17T17:26:31.657072536Z" level=error msg="encountered an error cleaning up failed sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.657233 containerd[2072]: time="2025-03-17T17:26:31.657172938Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:4,} failed, error" error="failed to setup network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.657502 kubelet[2554]: E0317 17:26:31.657444 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:31.657919 kubelet[2554]: E0317 17:26:31.657525 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:31.657919 kubelet[2554]: E0317 17:26:31.657560 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:31.657919 kubelet[2554]: E0317 17:26:31.657628 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:32.035719 kubelet[2554]: E0317 17:26:32.035287 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:32.293541 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c-shm.mount: Deactivated successfully. Mar 17 17:26:32.293934 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d-shm.mount: Deactivated successfully. Mar 17 17:26:32.400686 kubelet[2554]: I0317 17:26:32.400533 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c" Mar 17 17:26:32.402544 containerd[2072]: time="2025-03-17T17:26:32.401887140Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:26:32.402544 containerd[2072]: time="2025-03-17T17:26:32.402239548Z" level=info msg="Ensure that sandbox 021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c in task-service has been cleanup successfully" Mar 17 17:26:32.405225 containerd[2072]: time="2025-03-17T17:26:32.405144855Z" level=info msg="TearDown network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" successfully" Mar 17 17:26:32.405782 containerd[2072]: time="2025-03-17T17:26:32.405393382Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" returns successfully" Mar 17 17:26:32.409460 systemd[1]: run-netns-cni\x2d42061fb4\x2d8e37\x2d2e19\x2d6c67\x2dbff176a63b72.mount: Deactivated successfully. Mar 17 17:26:32.413345 containerd[2072]: time="2025-03-17T17:26:32.413276691Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:26:32.413501 containerd[2072]: time="2025-03-17T17:26:32.413453842Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:26:32.413501 containerd[2072]: time="2025-03-17T17:26:32.413484127Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:26:32.414509 containerd[2072]: time="2025-03-17T17:26:32.414430706Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:26:32.414661 containerd[2072]: time="2025-03-17T17:26:32.414600720Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:26:32.414661 containerd[2072]: time="2025-03-17T17:26:32.414629110Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:26:32.416405 containerd[2072]: time="2025-03-17T17:26:32.416136684Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:32.416405 containerd[2072]: time="2025-03-17T17:26:32.416305320Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:32.416405 containerd[2072]: time="2025-03-17T17:26:32.416334429Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:32.417366 containerd[2072]: time="2025-03-17T17:26:32.417099395Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:32.417366 containerd[2072]: time="2025-03-17T17:26:32.417298807Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:32.418785 containerd[2072]: time="2025-03-17T17:26:32.417327868Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:32.421149 containerd[2072]: time="2025-03-17T17:26:32.420988725Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:5,}" Mar 17 17:26:32.422796 kubelet[2554]: I0317 17:26:32.422464 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d" Mar 17 17:26:32.424490 containerd[2072]: time="2025-03-17T17:26:32.424412796Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:26:32.425758 containerd[2072]: time="2025-03-17T17:26:32.425453587Z" level=info msg="Ensure that sandbox ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d in task-service has been cleanup successfully" Mar 17 17:26:32.430209 systemd[1]: run-netns-cni\x2dccac707a\x2d84db\x2d2b78\x2dfeb1\x2d208d3ba78214.mount: Deactivated successfully. Mar 17 17:26:32.436527 containerd[2072]: time="2025-03-17T17:26:32.436263696Z" level=info msg="TearDown network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" successfully" Mar 17 17:26:32.436527 containerd[2072]: time="2025-03-17T17:26:32.436335900Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" returns successfully" Mar 17 17:26:32.442016 containerd[2072]: time="2025-03-17T17:26:32.441609493Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:26:32.442016 containerd[2072]: time="2025-03-17T17:26:32.441843101Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:26:32.442016 containerd[2072]: time="2025-03-17T17:26:32.441889110Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:26:32.444694 containerd[2072]: time="2025-03-17T17:26:32.444437307Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:26:32.446685 containerd[2072]: time="2025-03-17T17:26:32.446513192Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:26:32.447468 containerd[2072]: time="2025-03-17T17:26:32.447244610Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:26:32.451597 containerd[2072]: time="2025-03-17T17:26:32.451205005Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:32.451597 containerd[2072]: time="2025-03-17T17:26:32.451412046Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:32.451597 containerd[2072]: time="2025-03-17T17:26:32.451437797Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:32.452606 containerd[2072]: time="2025-03-17T17:26:32.452546990Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:32.454159 containerd[2072]: time="2025-03-17T17:26:32.454110972Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:32.454564 containerd[2072]: time="2025-03-17T17:26:32.454505514Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:32.456754 containerd[2072]: time="2025-03-17T17:26:32.456647223Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:32.456928 containerd[2072]: time="2025-03-17T17:26:32.456878143Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:32.456997 containerd[2072]: time="2025-03-17T17:26:32.456920926Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:32.458880 containerd[2072]: time="2025-03-17T17:26:32.458825249Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:6,}" Mar 17 17:26:32.661908 containerd[2072]: time="2025-03-17T17:26:32.661820653Z" level=error msg="Failed to destroy network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.667115 containerd[2072]: time="2025-03-17T17:26:32.666901888Z" level=error msg="encountered an error cleaning up failed sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.667806 containerd[2072]: time="2025-03-17T17:26:32.667347213Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:5,} failed, error" error="failed to setup network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.668162 kubelet[2554]: E0317 17:26:32.668093 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.668563 kubelet[2554]: E0317 17:26:32.668409 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:32.668563 kubelet[2554]: E0317 17:26:32.668531 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:32.668960 kubelet[2554]: E0317 17:26:32.668798 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:32.729860 containerd[2072]: time="2025-03-17T17:26:32.729792679Z" level=error msg="Failed to destroy network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.730707 containerd[2072]: time="2025-03-17T17:26:32.730639887Z" level=error msg="encountered an error cleaning up failed sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.730942 containerd[2072]: time="2025-03-17T17:26:32.730776870Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.732258 kubelet[2554]: E0317 17:26:32.731107 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:32.732258 kubelet[2554]: E0317 17:26:32.731204 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:32.732258 kubelet[2554]: E0317 17:26:32.731241 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:32.732516 kubelet[2554]: E0317 17:26:32.731335 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:33.024252 kubelet[2554]: E0317 17:26:33.023365 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:33.035877 kubelet[2554]: E0317 17:26:33.035828 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:33.226187 containerd[2072]: time="2025-03-17T17:26:33.226110313Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:33.227631 containerd[2072]: time="2025-03-17T17:26:33.227522643Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.2: active requests=0, bytes read=137086024" Mar 17 17:26:33.230774 containerd[2072]: time="2025-03-17T17:26:33.229511116Z" level=info msg="ImageCreate event name:\"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:33.235138 containerd[2072]: time="2025-03-17T17:26:33.235069315Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:33.238749 containerd[2072]: time="2025-03-17T17:26:33.238663833Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.2\" with image id \"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\", size \"137085886\" in 6.958736848s" Mar 17 17:26:33.238975 containerd[2072]: time="2025-03-17T17:26:33.238939587Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\" returns image reference \"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\"" Mar 17 17:26:33.253190 containerd[2072]: time="2025-03-17T17:26:33.253124459Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 17 17:26:33.271081 containerd[2072]: time="2025-03-17T17:26:33.270988515Z" level=info msg="CreateContainer within sandbox \"c52ec138c261f1dbb1695161adc39450755c3871b9a1acbbaffbdbcd5e70dc20\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"c9d0e01196430a70a2482644536ba3aafff2568f09b37730ed5bdabd43d5bc3c\"" Mar 17 17:26:33.272160 containerd[2072]: time="2025-03-17T17:26:33.272110206Z" level=info msg="StartContainer for \"c9d0e01196430a70a2482644536ba3aafff2568f09b37730ed5bdabd43d5bc3c\"" Mar 17 17:26:33.300247 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4-shm.mount: Deactivated successfully. Mar 17 17:26:33.300608 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2119942179.mount: Deactivated successfully. Mar 17 17:26:33.401537 containerd[2072]: time="2025-03-17T17:26:33.401356423Z" level=info msg="StartContainer for \"c9d0e01196430a70a2482644536ba3aafff2568f09b37730ed5bdabd43d5bc3c\" returns successfully" Mar 17 17:26:33.435889 kubelet[2554]: I0317 17:26:33.435218 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d" Mar 17 17:26:33.440208 containerd[2072]: time="2025-03-17T17:26:33.439184755Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" Mar 17 17:26:33.440208 containerd[2072]: time="2025-03-17T17:26:33.439483646Z" level=info msg="Ensure that sandbox d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d in task-service has been cleanup successfully" Mar 17 17:26:33.443054 containerd[2072]: time="2025-03-17T17:26:33.441871747Z" level=info msg="TearDown network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" successfully" Mar 17 17:26:33.443054 containerd[2072]: time="2025-03-17T17:26:33.442134560Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" returns successfully" Mar 17 17:26:33.443864 containerd[2072]: time="2025-03-17T17:26:33.443561474Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:26:33.444551 containerd[2072]: time="2025-03-17T17:26:33.444381060Z" level=info msg="TearDown network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" successfully" Mar 17 17:26:33.444849 containerd[2072]: time="2025-03-17T17:26:33.444665521Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" returns successfully" Mar 17 17:26:33.450297 containerd[2072]: time="2025-03-17T17:26:33.449632501Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:26:33.450297 containerd[2072]: time="2025-03-17T17:26:33.449873292Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:26:33.450297 containerd[2072]: time="2025-03-17T17:26:33.449901490Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:26:33.450620 systemd[1]: run-netns-cni\x2df6f37fb7\x2d32f7\x2d4913\x2d9519\x2da376f4a07897.mount: Deactivated successfully. Mar 17 17:26:33.454060 containerd[2072]: time="2025-03-17T17:26:33.452553460Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:26:33.454060 containerd[2072]: time="2025-03-17T17:26:33.452879708Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:26:33.454060 containerd[2072]: time="2025-03-17T17:26:33.452932098Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:26:33.454060 containerd[2072]: time="2025-03-17T17:26:33.453913795Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:33.455444 containerd[2072]: time="2025-03-17T17:26:33.455038005Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:33.455444 containerd[2072]: time="2025-03-17T17:26:33.455186982Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:33.455926 kubelet[2554]: I0317 17:26:33.455399 2554 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4" Mar 17 17:26:33.459365 containerd[2072]: time="2025-03-17T17:26:33.458441770Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" Mar 17 17:26:33.459365 containerd[2072]: time="2025-03-17T17:26:33.458794993Z" level=info msg="Ensure that sandbox ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4 in task-service has been cleanup successfully" Mar 17 17:26:33.460933 containerd[2072]: time="2025-03-17T17:26:33.460203137Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:33.460933 containerd[2072]: time="2025-03-17T17:26:33.460391143Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:33.460933 containerd[2072]: time="2025-03-17T17:26:33.460417242Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:33.462601 containerd[2072]: time="2025-03-17T17:26:33.462341211Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:33.470998 containerd[2072]: time="2025-03-17T17:26:33.463275460Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:33.471621 containerd[2072]: time="2025-03-17T17:26:33.471226774Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:33.471621 containerd[2072]: time="2025-03-17T17:26:33.463913373Z" level=info msg="TearDown network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" successfully" Mar 17 17:26:33.471621 containerd[2072]: time="2025-03-17T17:26:33.471498870Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" returns successfully" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.476147157Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.476329034Z" level=info msg="TearDown network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" successfully" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.476358011Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" returns successfully" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.476699373Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:7,}" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.477090233Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.477255870Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:26:33.479797 containerd[2072]: time="2025-03-17T17:26:33.477290137Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:26:33.478489 systemd[1]: run-netns-cni\x2d2778ca17\x2db298\x2dae10\x2d4ed7\x2ddd8bf9dfaf46.mount: Deactivated successfully. Mar 17 17:26:33.483339 containerd[2072]: time="2025-03-17T17:26:33.482986975Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:26:33.483715 containerd[2072]: time="2025-03-17T17:26:33.483295761Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:26:33.483715 containerd[2072]: time="2025-03-17T17:26:33.483460786Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:26:33.489524 containerd[2072]: time="2025-03-17T17:26:33.489043789Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:33.489524 containerd[2072]: time="2025-03-17T17:26:33.489247400Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:33.489524 containerd[2072]: time="2025-03-17T17:26:33.489278980Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:33.492147 containerd[2072]: time="2025-03-17T17:26:33.491482649Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:33.492147 containerd[2072]: time="2025-03-17T17:26:33.491671495Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:33.492147 containerd[2072]: time="2025-03-17T17:26:33.491707105Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:33.497058 containerd[2072]: time="2025-03-17T17:26:33.496431446Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:6,}" Mar 17 17:26:33.511984 kubelet[2554]: I0317 17:26:33.511560 2554 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-9qvxd" podStartSLOduration=3.718068826 podStartE2EDuration="20.511534183s" podCreationTimestamp="2025-03-17 17:26:13 +0000 UTC" firstStartedPulling="2025-03-17 17:26:16.446683082 +0000 UTC m=+4.365105154" lastFinishedPulling="2025-03-17 17:26:33.240148451 +0000 UTC m=+21.158570511" observedRunningTime="2025-03-17 17:26:33.5105243 +0000 UTC m=+21.428946420" watchObservedRunningTime="2025-03-17 17:26:33.511534183 +0000 UTC m=+21.429956255" Mar 17 17:26:33.563088 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Mar 17 17:26:33.563878 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.727 [INFO][3557] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.729 [INFO][3557] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" iface="eth0" netns="/var/run/netns/cni-96d06936-91cb-9772-cffc-6c6cc3bb69a2" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.729 [INFO][3557] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" iface="eth0" netns="/var/run/netns/cni-96d06936-91cb-9772-cffc-6c6cc3bb69a2" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.729 [INFO][3557] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" iface="eth0" netns="/var/run/netns/cni-96d06936-91cb-9772-cffc-6c6cc3bb69a2" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.729 [INFO][3557] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.729 [INFO][3557] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.808 [INFO][3573] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" HandleID="k8s-pod-network.d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.808 [INFO][3573] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.809 [INFO][3573] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.829 [WARNING][3573] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" HandleID="k8s-pod-network.d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.829 [INFO][3573] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" HandleID="k8s-pod-network.d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.832 [INFO][3573] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:26:33.839438 containerd[2072]: 2025-03-17 17:26:33.837 [INFO][3557] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09" Mar 17 17:26:33.844837 containerd[2072]: time="2025-03-17T17:26:33.844599995Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:6,} failed, error" error="failed to setup network for sandbox \"d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:33.850771 kubelet[2554]: E0317 17:26:33.847625 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:33.850771 kubelet[2554]: E0317 17:26:33.847757 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:33.850771 kubelet[2554]: E0317 17:26:33.847795 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="default/nginx-deployment-85f456d6dd-jshxr" Mar 17 17:26:33.851080 kubelet[2554]: E0317 17:26:33.847865 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nginx-deployment-85f456d6dd-jshxr_default(44cae95b-023f-401f-ab12-ec5a58221ec7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d17749a051136e0ff3ff730d1dfba7fd7f42e397293def9f8a293bc221438d09\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="default/nginx-deployment-85f456d6dd-jshxr" podUID="44cae95b-023f-401f-ab12-ec5a58221ec7" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.757 [INFO][3549] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.757 [INFO][3549] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" iface="eth0" netns="/var/run/netns/cni-e3a885a5-0f5a-81bb-2602-ed477dd68c1c" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.757 [INFO][3549] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" iface="eth0" netns="/var/run/netns/cni-e3a885a5-0f5a-81bb-2602-ed477dd68c1c" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.759 [INFO][3549] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" iface="eth0" netns="/var/run/netns/cni-e3a885a5-0f5a-81bb-2602-ed477dd68c1c" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.759 [INFO][3549] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.760 [INFO][3549] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.820 [INFO][3581] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" HandleID="k8s-pod-network.522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.820 [INFO][3581] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.832 [INFO][3581] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.847 [WARNING][3581] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" HandleID="k8s-pod-network.522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.847 [INFO][3581] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" HandleID="k8s-pod-network.522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.852 [INFO][3581] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:26:33.859574 containerd[2072]: 2025-03-17 17:26:33.856 [INFO][3549] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e" Mar 17 17:26:33.865871 containerd[2072]: time="2025-03-17T17:26:33.865790971Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:7,} failed, error" error="failed to setup network for sandbox \"522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:33.866285 kubelet[2554]: E0317 17:26:33.866219 2554 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:26:33.866420 kubelet[2554]: E0317 17:26:33.866319 2554 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:33.866420 kubelet[2554]: E0317 17:26:33.866354 2554 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-g9rjb" Mar 17 17:26:33.866535 kubelet[2554]: E0317 17:26:33.866418 2554 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-g9rjb_calico-system(d2efaf52-bf0a-463d-906a-d3ef9134a72b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"522fd16739ad6b43b59b58d94f76a42feccc2d7a02bd6438f581f8000780cb8e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-g9rjb" podUID="d2efaf52-bf0a-463d-906a-d3ef9134a72b" Mar 17 17:26:34.037182 kubelet[2554]: E0317 17:26:34.037078 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:34.485690 containerd[2072]: time="2025-03-17T17:26:34.485602967Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" Mar 17 17:26:34.486937 containerd[2072]: time="2025-03-17T17:26:34.485823417Z" level=info msg="TearDown network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" successfully" Mar 17 17:26:34.486937 containerd[2072]: time="2025-03-17T17:26:34.485873960Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" returns successfully" Mar 17 17:26:34.486937 containerd[2072]: time="2025-03-17T17:26:34.485617660Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" Mar 17 17:26:34.486937 containerd[2072]: time="2025-03-17T17:26:34.486147903Z" level=info msg="TearDown network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" successfully" Mar 17 17:26:34.486937 containerd[2072]: time="2025-03-17T17:26:34.486176029Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" returns successfully" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.486707519Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.487027591Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.487209396Z" level=info msg="TearDown network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" successfully" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.487366649Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" returns successfully" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.487465960Z" level=info msg="TearDown network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" successfully" Mar 17 17:26:34.487939 containerd[2072]: time="2025-03-17T17:26:34.487496028Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" returns successfully" Mar 17 17:26:34.488447 containerd[2072]: time="2025-03-17T17:26:34.488148562Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:26:34.488447 containerd[2072]: time="2025-03-17T17:26:34.488300970Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:26:34.488447 containerd[2072]: time="2025-03-17T17:26:34.488324046Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.488913887Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489089623Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489113875Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489349761Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489512880Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489536316Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:26:34.489630 containerd[2072]: time="2025-03-17T17:26:34.489584604Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:26:34.490126 containerd[2072]: time="2025-03-17T17:26:34.489715518Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:26:34.490126 containerd[2072]: time="2025-03-17T17:26:34.489773749Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:26:34.490926 containerd[2072]: time="2025-03-17T17:26:34.490860345Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:26:34.491092 containerd[2072]: time="2025-03-17T17:26:34.491051110Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:26:34.491189 containerd[2072]: time="2025-03-17T17:26:34.491089755Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:26:34.491689 containerd[2072]: time="2025-03-17T17:26:34.491389785Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:26:34.491689 containerd[2072]: time="2025-03-17T17:26:34.491566684Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:26:34.491689 containerd[2072]: time="2025-03-17T17:26:34.491591584Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:26:34.491689 containerd[2072]: time="2025-03-17T17:26:34.491679332Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:26:34.492005 containerd[2072]: time="2025-03-17T17:26:34.491927788Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:26:34.492005 containerd[2072]: time="2025-03-17T17:26:34.491981761Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:26:34.493137 containerd[2072]: time="2025-03-17T17:26:34.492799427Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:26:34.493137 containerd[2072]: time="2025-03-17T17:26:34.492816123Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:26:34.493137 containerd[2072]: time="2025-03-17T17:26:34.492984519Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:26:34.493137 containerd[2072]: time="2025-03-17T17:26:34.493008063Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:26:34.493709 containerd[2072]: time="2025-03-17T17:26:34.493649382Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:26:34.493709 containerd[2072]: time="2025-03-17T17:26:34.493698450Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:26:34.494540 containerd[2072]: time="2025-03-17T17:26:34.494483661Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:6,}" Mar 17 17:26:34.495375 containerd[2072]: time="2025-03-17T17:26:34.495301136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:7,}" Mar 17 17:26:34.747836 (udev-worker)[3513]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:26:34.749850 systemd-networkd[1604]: calie151de9358d: Link UP Mar 17 17:26:34.752074 systemd-networkd[1604]: calie151de9358d: Gained carrier Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.578 [INFO][3606] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.607 [INFO][3606] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.31.32-k8s-csi--node--driver--g9rjb-eth0 csi-node-driver- calico-system d2efaf52-bf0a-463d-906a-d3ef9134a72b 1116 0 2025-03-17 17:26:13 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:69ddf5d45d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s 172.31.31.32 csi-node-driver-g9rjb eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calie151de9358d [] []}} ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.607 [INFO][3606] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.674 [INFO][3625] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" HandleID="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.689 [INFO][3625] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" HandleID="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000319500), Attrs:map[string]string{"namespace":"calico-system", "node":"172.31.31.32", "pod":"csi-node-driver-g9rjb", "timestamp":"2025-03-17 17:26:34.67282156 +0000 UTC"}, Hostname:"172.31.31.32", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.690 [INFO][3625] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.690 [INFO][3625] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.690 [INFO][3625] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.31.32' Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.693 [INFO][3625] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.698 [INFO][3625] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.705 [INFO][3625] ipam/ipam.go 489: Trying affinity for 192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.708 [INFO][3625] ipam/ipam.go 155: Attempting to load block cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.711 [INFO][3625] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.711 [INFO][3625] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.27.192/26 handle="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.713 [INFO][3625] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.719 [INFO][3625] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.27.192/26 handle="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.729 [INFO][3625] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.27.193/26] block=192.168.27.192/26 handle="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.729 [INFO][3625] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.27.193/26] handle="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" host="172.31.31.32" Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.730 [INFO][3625] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:26:34.773233 containerd[2072]: 2025-03-17 17:26:34.730 [INFO][3625] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.27.193/26] IPv6=[] ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" HandleID="k8s-pod-network.31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Workload="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.733 [INFO][3606] cni-plugin/k8s.go 386: Populated endpoint ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-csi--node--driver--g9rjb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d2efaf52-bf0a-463d-906a-d3ef9134a72b", ResourceVersion:"1116", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"69ddf5d45d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"", Pod:"csi-node-driver-g9rjb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.27.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie151de9358d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.734 [INFO][3606] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.27.193/32] ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.734 [INFO][3606] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie151de9358d ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.754 [INFO][3606] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.755 [INFO][3606] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-csi--node--driver--g9rjb-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d2efaf52-bf0a-463d-906a-d3ef9134a72b", ResourceVersion:"1116", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"69ddf5d45d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd", Pod:"csi-node-driver-g9rjb", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.27.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie151de9358d", MAC:"d2:65:72:29:fe:e3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:34.774370 containerd[2072]: 2025-03-17 17:26:34.770 [INFO][3606] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd" Namespace="calico-system" Pod="csi-node-driver-g9rjb" WorkloadEndpoint="172.31.31.32-k8s-csi--node--driver--g9rjb-eth0" Mar 17 17:26:34.798102 systemd-networkd[1604]: cali1cc10a49012: Link UP Mar 17 17:26:34.799836 systemd-networkd[1604]: cali1cc10a49012: Gained carrier Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.584 [INFO][3600] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.609 [INFO][3600] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0 nginx-deployment-85f456d6dd- default 44cae95b-023f-401f-ab12-ec5a58221ec7 1115 0 2025-03-17 17:26:27 +0000 UTC map[app:nginx pod-template-hash:85f456d6dd projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.31.31.32 nginx-deployment-85f456d6dd-jshxr eth0 default [] [] [kns.default ksa.default.default] cali1cc10a49012 [] []}} ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.609 [INFO][3600] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.674 [INFO][3627] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" HandleID="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.697 [INFO][3627] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" HandleID="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400028d020), Attrs:map[string]string{"namespace":"default", "node":"172.31.31.32", "pod":"nginx-deployment-85f456d6dd-jshxr", "timestamp":"2025-03-17 17:26:34.67424057 +0000 UTC"}, Hostname:"172.31.31.32", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.697 [INFO][3627] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.730 [INFO][3627] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.730 [INFO][3627] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.31.32' Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.733 [INFO][3627] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.741 [INFO][3627] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.753 [INFO][3627] ipam/ipam.go 489: Trying affinity for 192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.758 [INFO][3627] ipam/ipam.go 155: Attempting to load block cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.762 [INFO][3627] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.762 [INFO][3627] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.27.192/26 handle="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.767 [INFO][3627] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34 Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.780 [INFO][3627] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.27.192/26 handle="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.787 [INFO][3627] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.27.194/26] block=192.168.27.192/26 handle="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.788 [INFO][3627] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.27.194/26] handle="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" host="172.31.31.32" Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.788 [INFO][3627] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:26:34.823387 containerd[2072]: 2025-03-17 17:26:34.788 [INFO][3627] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.27.194/26] IPv6=[] ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" HandleID="k8s-pod-network.4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Workload="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.790 [INFO][3600] cni-plugin/k8s.go 386: Populated endpoint ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0", GenerateName:"nginx-deployment-85f456d6dd-", Namespace:"default", SelfLink:"", UID:"44cae95b-023f-401f-ab12-ec5a58221ec7", ResourceVersion:"1115", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"85f456d6dd", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"", Pod:"nginx-deployment-85f456d6dd-jshxr", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.27.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali1cc10a49012", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.791 [INFO][3600] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.27.194/32] ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.791 [INFO][3600] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1cc10a49012 ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.801 [INFO][3600] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.803 [INFO][3600] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0", GenerateName:"nginx-deployment-85f456d6dd-", Namespace:"default", SelfLink:"", UID:"44cae95b-023f-401f-ab12-ec5a58221ec7", ResourceVersion:"1115", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 27, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nginx", "pod-template-hash":"85f456d6dd", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34", Pod:"nginx-deployment-85f456d6dd-jshxr", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.27.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali1cc10a49012", MAC:"d6:c7:be:d0:fd:2f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:34.824596 containerd[2072]: 2025-03-17 17:26:34.818 [INFO][3600] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34" Namespace="default" Pod="nginx-deployment-85f456d6dd-jshxr" WorkloadEndpoint="172.31.31.32-k8s-nginx--deployment--85f456d6dd--jshxr-eth0" Mar 17 17:26:34.830284 containerd[2072]: time="2025-03-17T17:26:34.830127944Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:26:34.830284 containerd[2072]: time="2025-03-17T17:26:34.830235866Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:26:34.830754 containerd[2072]: time="2025-03-17T17:26:34.830529275Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:34.831410 containerd[2072]: time="2025-03-17T17:26:34.831253749Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:34.880579 containerd[2072]: time="2025-03-17T17:26:34.880110672Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:26:34.880579 containerd[2072]: time="2025-03-17T17:26:34.880215344Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:26:34.880579 containerd[2072]: time="2025-03-17T17:26:34.880256087Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:34.881804 containerd[2072]: time="2025-03-17T17:26:34.880527452Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:34.928811 containerd[2072]: time="2025-03-17T17:26:34.928180777Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-g9rjb,Uid:d2efaf52-bf0a-463d-906a-d3ef9134a72b,Namespace:calico-system,Attempt:7,} returns sandbox id \"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd\"" Mar 17 17:26:34.948137 containerd[2072]: time="2025-03-17T17:26:34.948059429Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\"" Mar 17 17:26:34.994888 containerd[2072]: time="2025-03-17T17:26:34.994841079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-jshxr,Uid:44cae95b-023f-401f-ab12-ec5a58221ec7,Namespace:default,Attempt:6,} returns sandbox id \"4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34\"" Mar 17 17:26:35.038025 kubelet[2554]: E0317 17:26:35.037898 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:35.448774 kernel: bpftool[3860]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 17 17:26:35.735615 (udev-worker)[3509]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:26:35.736535 systemd-networkd[1604]: vxlan.calico: Link UP Mar 17 17:26:35.736556 systemd-networkd[1604]: vxlan.calico: Gained carrier Mar 17 17:26:36.040980 kubelet[2554]: E0317 17:26:36.040798 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:36.209328 systemd-networkd[1604]: cali1cc10a49012: Gained IPv6LL Mar 17 17:26:36.429509 containerd[2072]: time="2025-03-17T17:26:36.429418631Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:36.431274 containerd[2072]: time="2025-03-17T17:26:36.431164274Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.2: active requests=0, bytes read=7473801" Mar 17 17:26:36.432551 containerd[2072]: time="2025-03-17T17:26:36.432464891Z" level=info msg="ImageCreate event name:\"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:36.437784 containerd[2072]: time="2025-03-17T17:26:36.437674113Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:36.439599 containerd[2072]: time="2025-03-17T17:26:36.439337070Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.2\" with image id \"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\", size \"8843558\" in 1.491192184s" Mar 17 17:26:36.439599 containerd[2072]: time="2025-03-17T17:26:36.439405735Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\" returns image reference \"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\"" Mar 17 17:26:36.442480 containerd[2072]: time="2025-03-17T17:26:36.442170376Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Mar 17 17:26:36.443797 containerd[2072]: time="2025-03-17T17:26:36.443687521Z" level=info msg="CreateContainer within sandbox \"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 17 17:26:36.468026 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount58095092.mount: Deactivated successfully. Mar 17 17:26:36.473694 containerd[2072]: time="2025-03-17T17:26:36.473525175Z" level=info msg="CreateContainer within sandbox \"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"aa08132a0b13c338c0d4f0144142e382822866209d85dad8d3f9d0f9256b682d\"" Mar 17 17:26:36.475040 containerd[2072]: time="2025-03-17T17:26:36.474538716Z" level=info msg="StartContainer for \"aa08132a0b13c338c0d4f0144142e382822866209d85dad8d3f9d0f9256b682d\"" Mar 17 17:26:36.604234 containerd[2072]: time="2025-03-17T17:26:36.604019248Z" level=info msg="StartContainer for \"aa08132a0b13c338c0d4f0144142e382822866209d85dad8d3f9d0f9256b682d\" returns successfully" Mar 17 17:26:36.721090 systemd-networkd[1604]: calie151de9358d: Gained IPv6LL Mar 17 17:26:36.977465 systemd-networkd[1604]: vxlan.calico: Gained IPv6LL Mar 17 17:26:37.043787 kubelet[2554]: E0317 17:26:37.043655 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:37.266430 kubelet[2554]: I0317 17:26:37.266269 2554 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:26:38.044814 kubelet[2554]: E0317 17:26:38.044723 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:39.038034 ntpd[2011]: Listen normally on 6 vxlan.calico 192.168.27.192:123 Mar 17 17:26:39.038838 ntpd[2011]: 17 Mar 17:26:39 ntpd[2011]: Listen normally on 6 vxlan.calico 192.168.27.192:123 Mar 17 17:26:39.038838 ntpd[2011]: 17 Mar 17:26:39 ntpd[2011]: Listen normally on 7 calie151de9358d [fe80::ecee:eeff:feee:eeee%3]:123 Mar 17 17:26:39.038838 ntpd[2011]: 17 Mar 17:26:39 ntpd[2011]: Listen normally on 8 cali1cc10a49012 [fe80::ecee:eeff:feee:eeee%4]:123 Mar 17 17:26:39.038838 ntpd[2011]: 17 Mar 17:26:39 ntpd[2011]: Listen normally on 9 vxlan.calico [fe80::64ff:79ff:fe76:8e57%5]:123 Mar 17 17:26:39.038162 ntpd[2011]: Listen normally on 7 calie151de9358d [fe80::ecee:eeff:feee:eeee%3]:123 Mar 17 17:26:39.038240 ntpd[2011]: Listen normally on 8 cali1cc10a49012 [fe80::ecee:eeff:feee:eeee%4]:123 Mar 17 17:26:39.038307 ntpd[2011]: Listen normally on 9 vxlan.calico [fe80::64ff:79ff:fe76:8e57%5]:123 Mar 17 17:26:39.046938 kubelet[2554]: E0317 17:26:39.046853 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:40.047874 kubelet[2554]: E0317 17:26:40.047825 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:40.959540 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1786759929.mount: Deactivated successfully. Mar 17 17:26:41.048638 kubelet[2554]: E0317 17:26:41.048582 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:42.049637 kubelet[2554]: E0317 17:26:42.049593 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:42.457994 containerd[2072]: time="2025-03-17T17:26:42.457888087Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:42.460251 containerd[2072]: time="2025-03-17T17:26:42.460147204Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=69703867" Mar 17 17:26:42.462644 containerd[2072]: time="2025-03-17T17:26:42.462458052Z" level=info msg="ImageCreate event name:\"sha256:f660a383148a8217a75a455efeb8bfd4cbe3afa737712cc0e25f27c03b770dd4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:42.469181 containerd[2072]: time="2025-03-17T17:26:42.469069229Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/nginx@sha256:b927c62cc716b99bce51774b46a63feb63f5414c6f985fb80cacd1933bbd0e06\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:42.471544 containerd[2072]: time="2025-03-17T17:26:42.471283010Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:f660a383148a8217a75a455efeb8bfd4cbe3afa737712cc0e25f27c03b770dd4\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:b927c62cc716b99bce51774b46a63feb63f5414c6f985fb80cacd1933bbd0e06\", size \"69703745\" in 6.029054031s" Mar 17 17:26:42.471544 containerd[2072]: time="2025-03-17T17:26:42.471378746Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:f660a383148a8217a75a455efeb8bfd4cbe3afa737712cc0e25f27c03b770dd4\"" Mar 17 17:26:42.474554 containerd[2072]: time="2025-03-17T17:26:42.474454259Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\"" Mar 17 17:26:42.477464 containerd[2072]: time="2025-03-17T17:26:42.477114577Z" level=info msg="CreateContainer within sandbox \"4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Mar 17 17:26:42.511145 containerd[2072]: time="2025-03-17T17:26:42.511075995Z" level=info msg="CreateContainer within sandbox \"4110b133ab1976626e438ed4b5e5bab1070e5e1cefeb0f88386c45d57f340d34\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"9fd8d0607b4a48a5c9c6b810c8f10b76bf2fac776c75bb402b827bbb389e8c5f\"" Mar 17 17:26:42.512793 containerd[2072]: time="2025-03-17T17:26:42.512450256Z" level=info msg="StartContainer for \"9fd8d0607b4a48a5c9c6b810c8f10b76bf2fac776c75bb402b827bbb389e8c5f\"" Mar 17 17:26:42.631950 containerd[2072]: time="2025-03-17T17:26:42.631856775Z" level=info msg="StartContainer for \"9fd8d0607b4a48a5c9c6b810c8f10b76bf2fac776c75bb402b827bbb389e8c5f\" returns successfully" Mar 17 17:26:43.050538 kubelet[2554]: E0317 17:26:43.050455 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:44.050793 kubelet[2554]: E0317 17:26:44.050680 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:44.206947 containerd[2072]: time="2025-03-17T17:26:44.206880422Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:44.208889 containerd[2072]: time="2025-03-17T17:26:44.208777297Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2: active requests=0, bytes read=13121717" Mar 17 17:26:44.210248 containerd[2072]: time="2025-03-17T17:26:44.210161957Z" level=info msg="ImageCreate event name:\"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:44.214357 containerd[2072]: time="2025-03-17T17:26:44.214242927Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:44.216284 containerd[2072]: time="2025-03-17T17:26:44.215951568Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" with image id \"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\", size \"14491426\" in 1.741385596s" Mar 17 17:26:44.216284 containerd[2072]: time="2025-03-17T17:26:44.216012426Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" returns image reference \"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\"" Mar 17 17:26:44.220242 containerd[2072]: time="2025-03-17T17:26:44.220156089Z" level=info msg="CreateContainer within sandbox \"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 17 17:26:44.242803 containerd[2072]: time="2025-03-17T17:26:44.242114752Z" level=info msg="CreateContainer within sandbox \"31932798ccba3d27b3a0017345a6c72a3312b8fad4b350ff45a0928acf48fccd\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"c7a57880a38128e531d84043c4d7cf3d0dd8f9ae24f324f20e0f3e2dcc1a5b06\"" Mar 17 17:26:44.245202 containerd[2072]: time="2025-03-17T17:26:44.243764743Z" level=info msg="StartContainer for \"c7a57880a38128e531d84043c4d7cf3d0dd8f9ae24f324f20e0f3e2dcc1a5b06\"" Mar 17 17:26:44.303545 systemd[1]: run-containerd-runc-k8s.io-c7a57880a38128e531d84043c4d7cf3d0dd8f9ae24f324f20e0f3e2dcc1a5b06-runc.VWFry0.mount: Deactivated successfully. Mar 17 17:26:44.372933 containerd[2072]: time="2025-03-17T17:26:44.372777593Z" level=info msg="StartContainer for \"c7a57880a38128e531d84043c4d7cf3d0dd8f9ae24f324f20e0f3e2dcc1a5b06\" returns successfully" Mar 17 17:26:44.583212 update_engine[2032]: I20250317 17:26:44.583024 2032 update_attempter.cc:509] Updating boot flags... Mar 17 17:26:44.605211 kubelet[2554]: I0317 17:26:44.604682 2554 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-g9rjb" podStartSLOduration=22.332152827 podStartE2EDuration="31.604659034s" podCreationTimestamp="2025-03-17 17:26:13 +0000 UTC" firstStartedPulling="2025-03-17 17:26:34.945149493 +0000 UTC m=+22.863571577" lastFinishedPulling="2025-03-17 17:26:44.217655712 +0000 UTC m=+32.136077784" observedRunningTime="2025-03-17 17:26:44.604297032 +0000 UTC m=+32.522719200" watchObservedRunningTime="2025-03-17 17:26:44.604659034 +0000 UTC m=+32.523081178" Mar 17 17:26:44.605211 kubelet[2554]: I0317 17:26:44.604997 2554 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nginx-deployment-85f456d6dd-jshxr" podStartSLOduration=10.127813253 podStartE2EDuration="17.604980318s" podCreationTimestamp="2025-03-17 17:26:27 +0000 UTC" firstStartedPulling="2025-03-17 17:26:34.996966655 +0000 UTC m=+22.915388727" lastFinishedPulling="2025-03-17 17:26:42.474133708 +0000 UTC m=+30.392555792" observedRunningTime="2025-03-17 17:26:43.585713727 +0000 UTC m=+31.504135799" watchObservedRunningTime="2025-03-17 17:26:44.604980318 +0000 UTC m=+32.523402474" Mar 17 17:26:44.672932 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (4155) Mar 17 17:26:44.955980 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (4159) Mar 17 17:26:45.051312 kubelet[2554]: E0317 17:26:45.051216 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:45.171967 kubelet[2554]: I0317 17:26:45.170098 2554 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 17 17:26:45.171967 kubelet[2554]: I0317 17:26:45.170145 2554 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 17 17:26:45.218772 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (4159) Mar 17 17:26:46.051988 kubelet[2554]: E0317 17:26:46.051935 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:47.053396 kubelet[2554]: E0317 17:26:47.053320 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:48.054099 kubelet[2554]: E0317 17:26:48.054020 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:49.055056 kubelet[2554]: E0317 17:26:49.054979 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:49.595831 kubelet[2554]: I0317 17:26:49.595491 2554 topology_manager.go:215] "Topology Admit Handler" podUID="51aeeaa7-b983-428a-bc12-23000fde1d42" podNamespace="default" podName="nfs-server-provisioner-0" Mar 17 17:26:49.705680 kubelet[2554]: I0317 17:26:49.705546 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/51aeeaa7-b983-428a-bc12-23000fde1d42-data\") pod \"nfs-server-provisioner-0\" (UID: \"51aeeaa7-b983-428a-bc12-23000fde1d42\") " pod="default/nfs-server-provisioner-0" Mar 17 17:26:49.705680 kubelet[2554]: I0317 17:26:49.705622 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnz6r\" (UniqueName: \"kubernetes.io/projected/51aeeaa7-b983-428a-bc12-23000fde1d42-kube-api-access-rnz6r\") pod \"nfs-server-provisioner-0\" (UID: \"51aeeaa7-b983-428a-bc12-23000fde1d42\") " pod="default/nfs-server-provisioner-0" Mar 17 17:26:49.902960 containerd[2072]: time="2025-03-17T17:26:49.902874973Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:51aeeaa7-b983-428a-bc12-23000fde1d42,Namespace:default,Attempt:0,}" Mar 17 17:26:50.055842 kubelet[2554]: E0317 17:26:50.055767 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:50.117830 systemd-networkd[1604]: cali60e51b789ff: Link UP Mar 17 17:26:50.119505 systemd-networkd[1604]: cali60e51b789ff: Gained carrier Mar 17 17:26:50.126579 (udev-worker)[4443]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:49.987 [INFO][4422] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.31.32-k8s-nfs--server--provisioner--0-eth0 nfs-server-provisioner- default 51aeeaa7-b983-428a-bc12-23000fde1d42 1209 0 2025-03-17 17:26:49 +0000 UTC map[app:nfs-server-provisioner apps.kubernetes.io/pod-index:0 chart:nfs-server-provisioner-1.8.0 controller-revision-hash:nfs-server-provisioner-d5cbb7f57 heritage:Helm projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:nfs-server-provisioner release:nfs-server-provisioner statefulset.kubernetes.io/pod-name:nfs-server-provisioner-0] map[] [] [] []} {k8s 172.31.31.32 nfs-server-provisioner-0 eth0 nfs-server-provisioner [] [] [kns.default ksa.default.nfs-server-provisioner] cali60e51b789ff [{nfs TCP 2049 0 } {nfs-udp UDP 2049 0 } {nlockmgr TCP 32803 0 } {nlockmgr-udp UDP 32803 0 } {mountd TCP 20048 0 } {mountd-udp UDP 20048 0 } {rquotad TCP 875 0 } {rquotad-udp UDP 875 0 } {rpcbind TCP 111 0 } {rpcbind-udp UDP 111 0 } {statd TCP 662 0 } {statd-udp UDP 662 0 }] []}} ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:49.987 [INFO][4422] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.040 [INFO][4435] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" HandleID="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Workload="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.057 [INFO][4435] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" HandleID="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Workload="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002c39c0), Attrs:map[string]string{"namespace":"default", "node":"172.31.31.32", "pod":"nfs-server-provisioner-0", "timestamp":"2025-03-17 17:26:50.040023569 +0000 UTC"}, Hostname:"172.31.31.32", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.057 [INFO][4435] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.058 [INFO][4435] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.058 [INFO][4435] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.31.32' Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.060 [INFO][4435] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.067 [INFO][4435] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.075 [INFO][4435] ipam/ipam.go 489: Trying affinity for 192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.080 [INFO][4435] ipam/ipam.go 155: Attempting to load block cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.084 [INFO][4435] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.084 [INFO][4435] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.27.192/26 handle="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.086 [INFO][4435] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.093 [INFO][4435] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.27.192/26 handle="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.108 [INFO][4435] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.27.195/26] block=192.168.27.192/26 handle="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.108 [INFO][4435] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.27.195/26] handle="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" host="172.31.31.32" Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.108 [INFO][4435] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:26:50.135095 containerd[2072]: 2025-03-17 17:26:50.108 [INFO][4435] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.27.195/26] IPv6=[] ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" HandleID="k8s-pod-network.663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Workload="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.140377 containerd[2072]: 2025-03-17 17:26:50.111 [INFO][4422] cni-plugin/k8s.go 386: Populated endpoint ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"51aeeaa7-b983-428a-bc12-23000fde1d42", ResourceVersion:"1209", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.27.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:50.140377 containerd[2072]: 2025-03-17 17:26:50.111 [INFO][4422] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.27.195/32] ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.140377 containerd[2072]: 2025-03-17 17:26:50.111 [INFO][4422] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali60e51b789ff ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.140377 containerd[2072]: 2025-03-17 17:26:50.115 [INFO][4422] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.143848 containerd[2072]: 2025-03-17 17:26:50.115 [INFO][4422] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-nfs--server--provisioner--0-eth0", GenerateName:"nfs-server-provisioner-", Namespace:"default", SelfLink:"", UID:"51aeeaa7-b983-428a-bc12-23000fde1d42", ResourceVersion:"1209", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 49, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app":"nfs-server-provisioner", "apps.kubernetes.io/pod-index":"0", "chart":"nfs-server-provisioner-1.8.0", "controller-revision-hash":"nfs-server-provisioner-d5cbb7f57", "heritage":"Helm", "projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"nfs-server-provisioner", "release":"nfs-server-provisioner", "statefulset.kubernetes.io/pod-name":"nfs-server-provisioner-0"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba", Pod:"nfs-server-provisioner-0", Endpoint:"eth0", ServiceAccountName:"nfs-server-provisioner", IPNetworks:[]string{"192.168.27.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.nfs-server-provisioner"}, InterfaceName:"cali60e51b789ff", MAC:"96:32:37:44:90:29", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"nfs", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nfs-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x801, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"nlockmgr-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x8023, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"mountd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x4e50, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rquotad-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x36b, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"rpcbind-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x6f, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x296, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"statd-udp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x296, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:26:50.143848 containerd[2072]: 2025-03-17 17:26:50.132 [INFO][4422] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba" Namespace="default" Pod="nfs-server-provisioner-0" WorkloadEndpoint="172.31.31.32-k8s-nfs--server--provisioner--0-eth0" Mar 17 17:26:50.192538 containerd[2072]: time="2025-03-17T17:26:50.192023747Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:26:50.192661 containerd[2072]: time="2025-03-17T17:26:50.192388509Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:26:50.192661 containerd[2072]: time="2025-03-17T17:26:50.192447759Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:50.194154 containerd[2072]: time="2025-03-17T17:26:50.192874865Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:26:50.291821 containerd[2072]: time="2025-03-17T17:26:50.291757983Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:51aeeaa7-b983-428a-bc12-23000fde1d42,Namespace:default,Attempt:0,} returns sandbox id \"663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba\"" Mar 17 17:26:50.296015 containerd[2072]: time="2025-03-17T17:26:50.295551026Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Mar 17 17:26:51.057926 kubelet[2554]: E0317 17:26:51.057841 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:51.827668 systemd-networkd[1604]: cali60e51b789ff: Gained IPv6LL Mar 17 17:26:52.058871 kubelet[2554]: E0317 17:26:52.058579 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:52.960504 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount834811889.mount: Deactivated successfully. Mar 17 17:26:53.023265 kubelet[2554]: E0317 17:26:53.023206 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:53.058869 kubelet[2554]: E0317 17:26:53.058802 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:54.038554 ntpd[2011]: Listen normally on 10 cali60e51b789ff [fe80::ecee:eeff:feee:eeee%8]:123 Mar 17 17:26:54.040052 ntpd[2011]: 17 Mar 17:26:54 ntpd[2011]: Listen normally on 10 cali60e51b789ff [fe80::ecee:eeff:feee:eeee%8]:123 Mar 17 17:26:54.059336 kubelet[2554]: E0317 17:26:54.059256 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:55.061543 kubelet[2554]: E0317 17:26:55.061450 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:56.062194 kubelet[2554]: E0317 17:26:56.062141 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:56.066782 containerd[2072]: time="2025-03-17T17:26:56.066669480Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:56.069590 containerd[2072]: time="2025-03-17T17:26:56.069473726Z" level=info msg="stop pulling image registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8: active requests=0, bytes read=87373623" Mar 17 17:26:56.072343 containerd[2072]: time="2025-03-17T17:26:56.072225725Z" level=info msg="ImageCreate event name:\"sha256:5a42a519e0a8cf95c3c5f18f767c58c8c8b072aaea0a26e5e47a6f206c7df685\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:56.078802 containerd[2072]: time="2025-03-17T17:26:56.078711312Z" level=info msg="ImageCreate event name:\"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:26:56.081447 containerd[2072]: time="2025-03-17T17:26:56.081176583Z" level=info msg="Pulled image \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" with image id \"sha256:5a42a519e0a8cf95c3c5f18f767c58c8c8b072aaea0a26e5e47a6f206c7df685\", repo tag \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\", repo digest \"registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d\", size \"87371201\" in 5.785553869s" Mar 17 17:26:56.081447 containerd[2072]: time="2025-03-17T17:26:56.081261537Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:5a42a519e0a8cf95c3c5f18f767c58c8c8b072aaea0a26e5e47a6f206c7df685\"" Mar 17 17:26:56.087040 containerd[2072]: time="2025-03-17T17:26:56.086983610Z" level=info msg="CreateContainer within sandbox \"663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Mar 17 17:26:56.112653 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3267862958.mount: Deactivated successfully. Mar 17 17:26:56.118614 containerd[2072]: time="2025-03-17T17:26:56.118400946Z" level=info msg="CreateContainer within sandbox \"663c960f93587003e665deeed5d355c2a3705f7f997608c95d40b74b4a86b8ba\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"1a76f0bf442a4a6b301e1671d1bd75e1e7732ec258d2a28e99f12cccab30f037\"" Mar 17 17:26:56.119925 containerd[2072]: time="2025-03-17T17:26:56.119336562Z" level=info msg="StartContainer for \"1a76f0bf442a4a6b301e1671d1bd75e1e7732ec258d2a28e99f12cccab30f037\"" Mar 17 17:26:56.246911 containerd[2072]: time="2025-03-17T17:26:56.246575967Z" level=info msg="StartContainer for \"1a76f0bf442a4a6b301e1671d1bd75e1e7732ec258d2a28e99f12cccab30f037\" returns successfully" Mar 17 17:26:56.649505 kubelet[2554]: I0317 17:26:56.649357 2554 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=1.860336253 podStartE2EDuration="7.64933354s" podCreationTimestamp="2025-03-17 17:26:49 +0000 UTC" firstStartedPulling="2025-03-17 17:26:50.295070726 +0000 UTC m=+38.213492798" lastFinishedPulling="2025-03-17 17:26:56.084068013 +0000 UTC m=+44.002490085" observedRunningTime="2025-03-17 17:26:56.648005553 +0000 UTC m=+44.566427637" watchObservedRunningTime="2025-03-17 17:26:56.64933354 +0000 UTC m=+44.567755612" Mar 17 17:26:57.063451 kubelet[2554]: E0317 17:26:57.063276 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:58.063836 kubelet[2554]: E0317 17:26:58.063772 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:26:59.064482 kubelet[2554]: E0317 17:26:59.064399 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:00.065524 kubelet[2554]: E0317 17:27:00.065460 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:01.065834 kubelet[2554]: E0317 17:27:01.065773 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:02.066430 kubelet[2554]: E0317 17:27:02.066366 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:03.067423 kubelet[2554]: E0317 17:27:03.067359 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:04.067530 kubelet[2554]: E0317 17:27:04.067464 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:05.068608 kubelet[2554]: E0317 17:27:05.068548 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:06.069046 kubelet[2554]: E0317 17:27:06.068971 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:07.069856 kubelet[2554]: E0317 17:27:07.069781 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:08.070837 kubelet[2554]: E0317 17:27:08.070756 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:09.071333 kubelet[2554]: E0317 17:27:09.071267 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:10.072098 kubelet[2554]: E0317 17:27:10.072025 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:11.072757 kubelet[2554]: E0317 17:27:11.072661 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:12.073753 kubelet[2554]: E0317 17:27:12.073671 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:13.022855 kubelet[2554]: E0317 17:27:13.022799 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:13.049717 containerd[2072]: time="2025-03-17T17:27:13.049664109Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:27:13.050334 containerd[2072]: time="2025-03-17T17:27:13.049858400Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:27:13.050334 containerd[2072]: time="2025-03-17T17:27:13.049882544Z" level=info msg="StopPodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:27:13.050983 containerd[2072]: time="2025-03-17T17:27:13.050940175Z" level=info msg="RemovePodSandbox for \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:27:13.051079 containerd[2072]: time="2025-03-17T17:27:13.050990874Z" level=info msg="Forcibly stopping sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\"" Mar 17 17:27:13.051226 containerd[2072]: time="2025-03-17T17:27:13.051171983Z" level=info msg="TearDown network for sandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" successfully" Mar 17 17:27:13.059582 containerd[2072]: time="2025-03-17T17:27:13.059510620Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.059753 containerd[2072]: time="2025-03-17T17:27:13.059595225Z" level=info msg="RemovePodSandbox \"59f6cd76965070dad7038945698639da58e47c18d6093aed18bb04b8a00cc1fa\" returns successfully" Mar 17 17:27:13.060778 containerd[2072]: time="2025-03-17T17:27:13.060482554Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:27:13.060778 containerd[2072]: time="2025-03-17T17:27:13.060640359Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:27:13.060778 containerd[2072]: time="2025-03-17T17:27:13.060663627Z" level=info msg="StopPodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:27:13.061195 containerd[2072]: time="2025-03-17T17:27:13.061125756Z" level=info msg="RemovePodSandbox for \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:27:13.061195 containerd[2072]: time="2025-03-17T17:27:13.061176287Z" level=info msg="Forcibly stopping sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\"" Mar 17 17:27:13.061332 containerd[2072]: time="2025-03-17T17:27:13.061301192Z" level=info msg="TearDown network for sandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" successfully" Mar 17 17:27:13.072282 containerd[2072]: time="2025-03-17T17:27:13.071924362Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.072282 containerd[2072]: time="2025-03-17T17:27:13.072020410Z" level=info msg="RemovePodSandbox \"f131234997f369b3014c577055ee11d383e3afb2b71cbbb6a5783b19842bbbd8\" returns successfully" Mar 17 17:27:13.072996 containerd[2072]: time="2025-03-17T17:27:13.072562359Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:27:13.072996 containerd[2072]: time="2025-03-17T17:27:13.072710209Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:27:13.072996 containerd[2072]: time="2025-03-17T17:27:13.072764014Z" level=info msg="StopPodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:27:13.073925 kubelet[2554]: E0317 17:27:13.073886 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:13.074575 containerd[2072]: time="2025-03-17T17:27:13.074024931Z" level=info msg="RemovePodSandbox for \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:27:13.074575 containerd[2072]: time="2025-03-17T17:27:13.074074034Z" level=info msg="Forcibly stopping sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\"" Mar 17 17:27:13.074575 containerd[2072]: time="2025-03-17T17:27:13.074204865Z" level=info msg="TearDown network for sandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" successfully" Mar 17 17:27:13.079615 containerd[2072]: time="2025-03-17T17:27:13.079543958Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.079769 containerd[2072]: time="2025-03-17T17:27:13.079628660Z" level=info msg="RemovePodSandbox \"15eadc3ed3d0075c3af97e5f3b0a28dab6b52f5e368982938d730ebe079263c4\" returns successfully" Mar 17 17:27:13.080805 containerd[2072]: time="2025-03-17T17:27:13.080401001Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:27:13.080805 containerd[2072]: time="2025-03-17T17:27:13.080558399Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:27:13.080805 containerd[2072]: time="2025-03-17T17:27:13.080582111Z" level=info msg="StopPodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:27:13.081413 containerd[2072]: time="2025-03-17T17:27:13.081358626Z" level=info msg="RemovePodSandbox for \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:27:13.081490 containerd[2072]: time="2025-03-17T17:27:13.081410428Z" level=info msg="Forcibly stopping sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\"" Mar 17 17:27:13.081574 containerd[2072]: time="2025-03-17T17:27:13.081538237Z" level=info msg="TearDown network for sandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" successfully" Mar 17 17:27:13.087029 containerd[2072]: time="2025-03-17T17:27:13.086957713Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.087146 containerd[2072]: time="2025-03-17T17:27:13.087037390Z" level=info msg="RemovePodSandbox \"4b174ef8a5819e860958249c47fe95b3b1c3c50dc69974fab1b4524326130909\" returns successfully" Mar 17 17:27:13.087672 containerd[2072]: time="2025-03-17T17:27:13.087635446Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:27:13.088393 containerd[2072]: time="2025-03-17T17:27:13.088097335Z" level=info msg="TearDown network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" successfully" Mar 17 17:27:13.088393 containerd[2072]: time="2025-03-17T17:27:13.088127056Z" level=info msg="StopPodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" returns successfully" Mar 17 17:27:13.089201 containerd[2072]: time="2025-03-17T17:27:13.088913791Z" level=info msg="RemovePodSandbox for \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:27:13.089201 containerd[2072]: time="2025-03-17T17:27:13.088973593Z" level=info msg="Forcibly stopping sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\"" Mar 17 17:27:13.089201 containerd[2072]: time="2025-03-17T17:27:13.089094756Z" level=info msg="TearDown network for sandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" successfully" Mar 17 17:27:13.094515 containerd[2072]: time="2025-03-17T17:27:13.094448398Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.094642 containerd[2072]: time="2025-03-17T17:27:13.094530137Z" level=info msg="RemovePodSandbox \"021aec6a511281b1f59392dd675ec6e8bf267e2cf96fd50d3e8d78ef5fbd141c\" returns successfully" Mar 17 17:27:13.095177 containerd[2072]: time="2025-03-17T17:27:13.095128122Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" Mar 17 17:27:13.095594 containerd[2072]: time="2025-03-17T17:27:13.095284704Z" level=info msg="TearDown network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" successfully" Mar 17 17:27:13.095594 containerd[2072]: time="2025-03-17T17:27:13.095324356Z" level=info msg="StopPodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" returns successfully" Mar 17 17:27:13.096793 containerd[2072]: time="2025-03-17T17:27:13.095976470Z" level=info msg="RemovePodSandbox for \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" Mar 17 17:27:13.096793 containerd[2072]: time="2025-03-17T17:27:13.096025705Z" level=info msg="Forcibly stopping sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\"" Mar 17 17:27:13.096793 containerd[2072]: time="2025-03-17T17:27:13.096144302Z" level=info msg="TearDown network for sandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" successfully" Mar 17 17:27:13.101410 containerd[2072]: time="2025-03-17T17:27:13.101346963Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.101617 containerd[2072]: time="2025-03-17T17:27:13.101428654Z" level=info msg="RemovePodSandbox \"ab18ec3af1c36399252f0ca1b6613c689ef450661dc0012dde27fc7d01e96cf4\" returns successfully" Mar 17 17:27:13.102596 containerd[2072]: time="2025-03-17T17:27:13.102147587Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:27:13.102596 containerd[2072]: time="2025-03-17T17:27:13.102300414Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:27:13.102596 containerd[2072]: time="2025-03-17T17:27:13.102321512Z" level=info msg="StopPodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:27:13.103454 containerd[2072]: time="2025-03-17T17:27:13.103163718Z" level=info msg="RemovePodSandbox for \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:27:13.103454 containerd[2072]: time="2025-03-17T17:27:13.103219874Z" level=info msg="Forcibly stopping sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\"" Mar 17 17:27:13.103454 containerd[2072]: time="2025-03-17T17:27:13.103346135Z" level=info msg="TearDown network for sandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" successfully" Mar 17 17:27:13.109952 containerd[2072]: time="2025-03-17T17:27:13.109887639Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.110419 containerd[2072]: time="2025-03-17T17:27:13.109978481Z" level=info msg="RemovePodSandbox \"be0274dfcefed54586ef6b6d57d85ea45a73435d307f9c23f39e61af8243b820\" returns successfully" Mar 17 17:27:13.111238 containerd[2072]: time="2025-03-17T17:27:13.110684904Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:27:13.111238 containerd[2072]: time="2025-03-17T17:27:13.110869396Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:27:13.111238 containerd[2072]: time="2025-03-17T17:27:13.110892796Z" level=info msg="StopPodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:27:13.112250 containerd[2072]: time="2025-03-17T17:27:13.111991554Z" level=info msg="RemovePodSandbox for \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:27:13.112250 containerd[2072]: time="2025-03-17T17:27:13.112041629Z" level=info msg="Forcibly stopping sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\"" Mar 17 17:27:13.112250 containerd[2072]: time="2025-03-17T17:27:13.112173647Z" level=info msg="TearDown network for sandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" successfully" Mar 17 17:27:13.118276 containerd[2072]: time="2025-03-17T17:27:13.118055817Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.118276 containerd[2072]: time="2025-03-17T17:27:13.118145016Z" level=info msg="RemovePodSandbox \"a8e58a5446aa73fecac74e0cf4834eefc856c3fe0b1b68435ef9edeb1951107e\" returns successfully" Mar 17 17:27:13.119246 containerd[2072]: time="2025-03-17T17:27:13.118951529Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:27:13.119246 containerd[2072]: time="2025-03-17T17:27:13.119121688Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:27:13.119246 containerd[2072]: time="2025-03-17T17:27:13.119143553Z" level=info msg="StopPodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:27:13.119968 containerd[2072]: time="2025-03-17T17:27:13.119910929Z" level=info msg="RemovePodSandbox for \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:27:13.120076 containerd[2072]: time="2025-03-17T17:27:13.119967456Z" level=info msg="Forcibly stopping sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\"" Mar 17 17:27:13.120136 containerd[2072]: time="2025-03-17T17:27:13.120094737Z" level=info msg="TearDown network for sandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" successfully" Mar 17 17:27:13.125525 containerd[2072]: time="2025-03-17T17:27:13.125452505Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.125642 containerd[2072]: time="2025-03-17T17:27:13.125534124Z" level=info msg="RemovePodSandbox \"882c9e188cb83d6b894417a59472c0f75040a55cac9f5475adc317a032503a24\" returns successfully" Mar 17 17:27:13.126375 containerd[2072]: time="2025-03-17T17:27:13.126336067Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:27:13.126847 containerd[2072]: time="2025-03-17T17:27:13.126687947Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:27:13.126847 containerd[2072]: time="2025-03-17T17:27:13.126715725Z" level=info msg="StopPodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:27:13.127609 containerd[2072]: time="2025-03-17T17:27:13.127508576Z" level=info msg="RemovePodSandbox for \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:27:13.127609 containerd[2072]: time="2025-03-17T17:27:13.127552474Z" level=info msg="Forcibly stopping sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\"" Mar 17 17:27:13.127794 containerd[2072]: time="2025-03-17T17:27:13.127695191Z" level=info msg="TearDown network for sandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" successfully" Mar 17 17:27:13.133023 containerd[2072]: time="2025-03-17T17:27:13.132958590Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.133196 containerd[2072]: time="2025-03-17T17:27:13.133037342Z" level=info msg="RemovePodSandbox \"43f109b5cfbab644b5eac60c60a05e7213b8ca41363346a7f993f6613c5c58b6\" returns successfully" Mar 17 17:27:13.134077 containerd[2072]: time="2025-03-17T17:27:13.133829498Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:27:13.134077 containerd[2072]: time="2025-03-17T17:27:13.133983201Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:27:13.134077 containerd[2072]: time="2025-03-17T17:27:13.134004574Z" level=info msg="StopPodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:27:13.134988 containerd[2072]: time="2025-03-17T17:27:13.134947974Z" level=info msg="RemovePodSandbox for \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:27:13.135115 containerd[2072]: time="2025-03-17T17:27:13.134994067Z" level=info msg="Forcibly stopping sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\"" Mar 17 17:27:13.135267 containerd[2072]: time="2025-03-17T17:27:13.135112712Z" level=info msg="TearDown network for sandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" successfully" Mar 17 17:27:13.140652 containerd[2072]: time="2025-03-17T17:27:13.140582048Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.140802 containerd[2072]: time="2025-03-17T17:27:13.140660321Z" level=info msg="RemovePodSandbox \"8c239d36930d0419158d063b136d7ed5378ac59d3805a3b039efeb5d0d103792\" returns successfully" Mar 17 17:27:13.141373 containerd[2072]: time="2025-03-17T17:27:13.141323385Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:27:13.141540 containerd[2072]: time="2025-03-17T17:27:13.141497250Z" level=info msg="TearDown network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" successfully" Mar 17 17:27:13.141607 containerd[2072]: time="2025-03-17T17:27:13.141535943Z" level=info msg="StopPodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" returns successfully" Mar 17 17:27:13.142581 containerd[2072]: time="2025-03-17T17:27:13.142427721Z" level=info msg="RemovePodSandbox for \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:27:13.142581 containerd[2072]: time="2025-03-17T17:27:13.142496686Z" level=info msg="Forcibly stopping sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\"" Mar 17 17:27:13.143317 containerd[2072]: time="2025-03-17T17:27:13.143066785Z" level=info msg="TearDown network for sandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" successfully" Mar 17 17:27:13.149590 containerd[2072]: time="2025-03-17T17:27:13.149360673Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.149590 containerd[2072]: time="2025-03-17T17:27:13.149441332Z" level=info msg="RemovePodSandbox \"ede59bd06e7c57367c8cf158d0241a68a46a86cd462cc73b7e5fd628dd62e74d\" returns successfully" Mar 17 17:27:13.150784 containerd[2072]: time="2025-03-17T17:27:13.150279653Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" Mar 17 17:27:13.150784 containerd[2072]: time="2025-03-17T17:27:13.150439137Z" level=info msg="TearDown network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" successfully" Mar 17 17:27:13.150784 containerd[2072]: time="2025-03-17T17:27:13.150467323Z" level=info msg="StopPodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" returns successfully" Mar 17 17:27:13.151788 containerd[2072]: time="2025-03-17T17:27:13.151212426Z" level=info msg="RemovePodSandbox for \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" Mar 17 17:27:13.151788 containerd[2072]: time="2025-03-17T17:27:13.151260882Z" level=info msg="Forcibly stopping sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\"" Mar 17 17:27:13.151788 containerd[2072]: time="2025-03-17T17:27:13.151425644Z" level=info msg="TearDown network for sandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" successfully" Mar 17 17:27:13.186529 containerd[2072]: time="2025-03-17T17:27:13.186451470Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:27:13.186684 containerd[2072]: time="2025-03-17T17:27:13.186547086Z" level=info msg="RemovePodSandbox \"d1af0f5ff5f097d05d780b5f6eca8a7ac6cd8e469670c45acce38cfcdc81eb5d\" returns successfully" Mar 17 17:27:14.074397 kubelet[2554]: E0317 17:27:14.074333 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:15.075064 kubelet[2554]: E0317 17:27:15.075000 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:16.075466 kubelet[2554]: E0317 17:27:16.075393 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:17.075872 kubelet[2554]: E0317 17:27:17.075805 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:18.076469 kubelet[2554]: E0317 17:27:18.076416 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:19.077169 kubelet[2554]: E0317 17:27:19.077086 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:20.077916 kubelet[2554]: E0317 17:27:20.077854 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:21.078233 kubelet[2554]: E0317 17:27:21.078169 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:21.445638 kubelet[2554]: I0317 17:27:21.445573 2554 topology_manager.go:215] "Topology Admit Handler" podUID="d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4" podNamespace="default" podName="test-pod-1" Mar 17 17:27:21.521750 kubelet[2554]: I0317 17:27:21.521568 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-60d308a7-9128-4960-b062-cd5f3237aceb\" (UniqueName: \"kubernetes.io/nfs/d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4-pvc-60d308a7-9128-4960-b062-cd5f3237aceb\") pod \"test-pod-1\" (UID: \"d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4\") " pod="default/test-pod-1" Mar 17 17:27:21.521750 kubelet[2554]: I0317 17:27:21.521633 2554 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmhtb\" (UniqueName: \"kubernetes.io/projected/d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4-kube-api-access-pmhtb\") pod \"test-pod-1\" (UID: \"d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4\") " pod="default/test-pod-1" Mar 17 17:27:21.657885 kernel: FS-Cache: Loaded Mar 17 17:27:21.700502 kernel: RPC: Registered named UNIX socket transport module. Mar 17 17:27:21.700652 kernel: RPC: Registered udp transport module. Mar 17 17:27:21.700696 kernel: RPC: Registered tcp transport module. Mar 17 17:27:21.702583 kernel: RPC: Registered tcp-with-tls transport module. Mar 17 17:27:21.702757 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Mar 17 17:27:22.037776 kernel: NFS: Registering the id_resolver key type Mar 17 17:27:22.037909 kernel: Key type id_resolver registered Mar 17 17:27:22.037952 kernel: Key type id_legacy registered Mar 17 17:27:22.078879 kubelet[2554]: E0317 17:27:22.078827 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:22.080061 nfsidmap[4663]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' Mar 17 17:27:22.087557 nfsidmap[4664]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' Mar 17 17:27:22.354251 containerd[2072]: time="2025-03-17T17:27:22.354177757Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4,Namespace:default,Attempt:0,}" Mar 17 17:27:22.554465 (udev-worker)[4650]: Network interface NamePolicy= disabled on kernel command line. Mar 17 17:27:22.554574 systemd-networkd[1604]: cali5ec59c6bf6e: Link UP Mar 17 17:27:22.555463 systemd-networkd[1604]: cali5ec59c6bf6e: Gained carrier Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.432 [INFO][4666] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {172.31.31.32-k8s-test--pod--1-eth0 default d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4 1332 0 2025-03-17 17:26:50 +0000 UTC map[projectcalico.org/namespace:default projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s 172.31.31.32 test-pod-1 eth0 default [] [] [kns.default ksa.default.default] cali5ec59c6bf6e [] []}} ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.433 [INFO][4666] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.479 [INFO][4677] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" HandleID="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Workload="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.499 [INFO][4677] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" HandleID="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Workload="172.31.31.32-k8s-test--pod--1-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000221110), Attrs:map[string]string{"namespace":"default", "node":"172.31.31.32", "pod":"test-pod-1", "timestamp":"2025-03-17 17:27:22.479664395 +0000 UTC"}, Hostname:"172.31.31.32", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.499 [INFO][4677] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.499 [INFO][4677] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.499 [INFO][4677] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host '172.31.31.32' Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.502 [INFO][4677] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.509 [INFO][4677] ipam/ipam.go 372: Looking up existing affinities for host host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.516 [INFO][4677] ipam/ipam.go 489: Trying affinity for 192.168.27.192/26 host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.520 [INFO][4677] ipam/ipam.go 155: Attempting to load block cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.524 [INFO][4677] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.27.192/26 host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.524 [INFO][4677] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.27.192/26 handle="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.527 [INFO][4677] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419 Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.534 [INFO][4677] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.27.192/26 handle="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.546 [INFO][4677] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.27.196/26] block=192.168.27.192/26 handle="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.546 [INFO][4677] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.27.196/26] handle="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" host="172.31.31.32" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.546 [INFO][4677] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.546 [INFO][4677] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.27.196/26] IPv6=[] ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" HandleID="k8s-pod-network.49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Workload="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.577854 containerd[2072]: 2025-03-17 17:27:22.550 [INFO][4666] cni-plugin/k8s.go 386: Populated endpoint ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4", ResourceVersion:"1332", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.27.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:27:22.579924 containerd[2072]: 2025-03-17 17:27:22.550 [INFO][4666] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.27.196/32] ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.579924 containerd[2072]: 2025-03-17 17:27:22.550 [INFO][4666] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5ec59c6bf6e ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.579924 containerd[2072]: 2025-03-17 17:27:22.555 [INFO][4666] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.579924 containerd[2072]: 2025-03-17 17:27:22.557 [INFO][4666] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"172.31.31.32-k8s-test--pod--1-eth0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4", ResourceVersion:"1332", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 26, 50, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"projectcalico.org/namespace":"default", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"172.31.31.32", ContainerID:"49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419", Pod:"test-pod-1", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.27.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.default", "ksa.default.default"}, InterfaceName:"cali5ec59c6bf6e", MAC:"8a:49:9e:09:1c:4d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:27:22.579924 containerd[2072]: 2025-03-17 17:27:22.570 [INFO][4666] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419" Namespace="default" Pod="test-pod-1" WorkloadEndpoint="172.31.31.32-k8s-test--pod--1-eth0" Mar 17 17:27:22.615790 containerd[2072]: time="2025-03-17T17:27:22.615173447Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:27:22.615790 containerd[2072]: time="2025-03-17T17:27:22.615284475Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:27:22.615790 containerd[2072]: time="2025-03-17T17:27:22.615412223Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:27:22.617176 containerd[2072]: time="2025-03-17T17:27:22.616505836Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:27:22.706437 containerd[2072]: time="2025-03-17T17:27:22.706383871Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:d90191a8-e2e0-4e3d-956e-9ea1d2fca9a4,Namespace:default,Attempt:0,} returns sandbox id \"49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419\"" Mar 17 17:27:22.709214 containerd[2072]: time="2025-03-17T17:27:22.709157927Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Mar 17 17:27:23.080927 kubelet[2554]: E0317 17:27:23.080706 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:23.119168 containerd[2072]: time="2025-03-17T17:27:23.119076255Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/nginx:latest\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:27:23.122860 containerd[2072]: time="2025-03-17T17:27:23.122709861Z" level=info msg="stop pulling image ghcr.io/flatcar/nginx:latest: active requests=0, bytes read=61" Mar 17 17:27:23.130150 containerd[2072]: time="2025-03-17T17:27:23.130071550Z" level=info msg="Pulled image \"ghcr.io/flatcar/nginx:latest\" with image id \"sha256:f660a383148a8217a75a455efeb8bfd4cbe3afa737712cc0e25f27c03b770dd4\", repo tag \"ghcr.io/flatcar/nginx:latest\", repo digest \"ghcr.io/flatcar/nginx@sha256:b927c62cc716b99bce51774b46a63feb63f5414c6f985fb80cacd1933bbd0e06\", size \"69703745\" in 420.853893ms" Mar 17 17:27:23.130150 containerd[2072]: time="2025-03-17T17:27:23.130137169Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:f660a383148a8217a75a455efeb8bfd4cbe3afa737712cc0e25f27c03b770dd4\"" Mar 17 17:27:23.135854 containerd[2072]: time="2025-03-17T17:27:23.135486422Z" level=info msg="CreateContainer within sandbox \"49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419\" for container &ContainerMetadata{Name:test,Attempt:0,}" Mar 17 17:27:23.165118 containerd[2072]: time="2025-03-17T17:27:23.165027572Z" level=info msg="CreateContainer within sandbox \"49f752886f729c22cf09accd72454328d65c7a69eadae479f1dfcdb58fd98419\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"8728a2cf0409195f938558b42a55469e222bcaf06d18da44abca1422d88d6dec\"" Mar 17 17:27:23.167488 containerd[2072]: time="2025-03-17T17:27:23.167382150Z" level=info msg="StartContainer for \"8728a2cf0409195f938558b42a55469e222bcaf06d18da44abca1422d88d6dec\"" Mar 17 17:27:23.281447 containerd[2072]: time="2025-03-17T17:27:23.281098631Z" level=info msg="StartContainer for \"8728a2cf0409195f938558b42a55469e222bcaf06d18da44abca1422d88d6dec\" returns successfully" Mar 17 17:27:23.889023 systemd-networkd[1604]: cali5ec59c6bf6e: Gained IPv6LL Mar 17 17:27:24.081579 kubelet[2554]: E0317 17:27:24.081483 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:25.081758 kubelet[2554]: E0317 17:27:25.081660 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:26.038072 ntpd[2011]: Listen normally on 11 cali5ec59c6bf6e [fe80::ecee:eeff:feee:eeee%9]:123 Mar 17 17:27:26.038667 ntpd[2011]: 17 Mar 17:27:26 ntpd[2011]: Listen normally on 11 cali5ec59c6bf6e [fe80::ecee:eeff:feee:eeee%9]:123 Mar 17 17:27:26.082743 kubelet[2554]: E0317 17:27:26.082687 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:27.083263 kubelet[2554]: E0317 17:27:27.083195 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:28.083816 kubelet[2554]: E0317 17:27:28.083717 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:29.084252 kubelet[2554]: E0317 17:27:29.084186 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:30.084532 kubelet[2554]: E0317 17:27:30.084460 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:31.084959 kubelet[2554]: E0317 17:27:31.084884 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:32.085554 kubelet[2554]: E0317 17:27:32.085483 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:33.023568 kubelet[2554]: E0317 17:27:33.023502 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:33.085883 kubelet[2554]: E0317 17:27:33.085821 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:34.086644 kubelet[2554]: E0317 17:27:34.086570 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:35.086959 kubelet[2554]: E0317 17:27:35.086896 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:36.087322 kubelet[2554]: E0317 17:27:36.087251 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:37.088400 kubelet[2554]: E0317 17:27:37.088328 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:38.089185 kubelet[2554]: E0317 17:27:38.089117 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:39.090090 kubelet[2554]: E0317 17:27:39.090017 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:40.090723 kubelet[2554]: E0317 17:27:40.090649 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:41.091780 kubelet[2554]: E0317 17:27:41.091685 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:42.092162 kubelet[2554]: E0317 17:27:42.092091 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:43.093123 kubelet[2554]: E0317 17:27:43.093074 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:44.093569 kubelet[2554]: E0317 17:27:44.093510 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:45.035004 kubelet[2554]: E0317 17:27:45.034929 2554 controller.go:195] "Failed to update lease" err="the server was unable to return a response in the time allotted, but may still be processing the request (put leases.coordination.k8s.io 172.31.31.32)" Mar 17 17:27:45.094650 kubelet[2554]: E0317 17:27:45.094578 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:46.095265 kubelet[2554]: E0317 17:27:46.095178 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:47.095973 kubelet[2554]: E0317 17:27:47.095896 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:48.096119 kubelet[2554]: E0317 17:27:48.096048 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:49.096625 kubelet[2554]: E0317 17:27:49.096547 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:50.097579 kubelet[2554]: E0317 17:27:50.097505 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:51.098410 kubelet[2554]: E0317 17:27:51.098339 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:52.099335 kubelet[2554]: E0317 17:27:52.099269 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:53.023249 kubelet[2554]: E0317 17:27:53.023194 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:53.100075 kubelet[2554]: E0317 17:27:53.099995 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:54.100397 kubelet[2554]: E0317 17:27:54.100323 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:55.031435 kubelet[2554]: E0317 17:27:55.031231 2554 controller.go:195] "Failed to update lease" err="the server was unable to return a response in the time allotted, but may still be processing the request (put leases.coordination.k8s.io 172.31.31.32)" Mar 17 17:27:55.101254 kubelet[2554]: E0317 17:27:55.101169 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:56.101446 kubelet[2554]: E0317 17:27:56.101374 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:57.102269 kubelet[2554]: E0317 17:27:57.102203 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:58.103037 kubelet[2554]: E0317 17:27:58.102975 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:27:59.103753 kubelet[2554]: E0317 17:27:59.103675 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:00.104487 kubelet[2554]: E0317 17:28:00.104418 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:01.105478 kubelet[2554]: E0317 17:28:01.105414 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:02.106039 kubelet[2554]: E0317 17:28:02.105974 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:03.107131 kubelet[2554]: E0317 17:28:03.107061 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:04.107704 kubelet[2554]: E0317 17:28:04.107640 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:05.028151 kubelet[2554]: E0317 17:28:05.027827 2554 controller.go:195] "Failed to update lease" err="the server was unable to return a response in the time allotted, but may still be processing the request (put leases.coordination.k8s.io 172.31.31.32)" Mar 17 17:28:05.108144 kubelet[2554]: E0317 17:28:05.108103 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:06.109790 kubelet[2554]: E0317 17:28:06.109706 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:07.110473 kubelet[2554]: E0317 17:28:07.110402 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:08.111562 kubelet[2554]: E0317 17:28:08.111496 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:08.765834 kubelet[2554]: E0317 17:28:08.765754 2554 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": unexpected EOF" Mar 17 17:28:08.778703 kubelet[2554]: E0317 17:28:08.777620 2554 controller.go:195] "Failed to update lease" err="Put \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": read tcp 172.31.31.32:51538->172.31.24.234:6443: read: connection reset by peer" Mar 17 17:28:08.778703 kubelet[2554]: I0317 17:28:08.777674 2554 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 17 17:28:08.778703 kubelet[2554]: E0317 17:28:08.778159 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": dial tcp 172.31.24.234:6443: connect: connection refused" interval="200ms" Mar 17 17:28:08.979323 kubelet[2554]: E0317 17:28:08.979257 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": dial tcp 172.31.24.234:6443: connect: connection refused" interval="400ms" Mar 17 17:28:09.112287 kubelet[2554]: E0317 17:28:09.112217 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:09.380488 kubelet[2554]: E0317 17:28:09.380339 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": dial tcp 172.31.24.234:6443: connect: connection refused" interval="800ms" Mar 17 17:28:10.112678 kubelet[2554]: E0317 17:28:10.112589 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:10.181369 kubelet[2554]: E0317 17:28:10.181308 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": dial tcp 172.31.24.234:6443: connect: connection refused" interval="1.6s" Mar 17 17:28:11.112877 kubelet[2554]: E0317 17:28:11.112809 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:11.782710 kubelet[2554]: E0317 17:28:11.782630 2554 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.24.234:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.31.32?timeout=10s\": dial tcp 172.31.24.234:6443: connect: connection refused" interval="3.2s" Mar 17 17:28:12.114067 kubelet[2554]: E0317 17:28:12.113979 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:13.023041 kubelet[2554]: E0317 17:28:13.022975 2554 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:13.114820 kubelet[2554]: E0317 17:28:13.114708 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:14.115487 kubelet[2554]: E0317 17:28:14.115424 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:15.116150 kubelet[2554]: E0317 17:28:15.116089 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:16.117222 kubelet[2554]: E0317 17:28:16.117158 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:17.118183 kubelet[2554]: E0317 17:28:17.118119 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:18.119232 kubelet[2554]: E0317 17:28:18.119174 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:19.119789 kubelet[2554]: E0317 17:28:19.119699 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Mar 17 17:28:20.120611 kubelet[2554]: E0317 17:28:20.120542 2554 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"